code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _lowerCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( lowerCAmelCase_ ): def __init__( self , *__snake_case , **__snake_case ) -> None: '''simple docstring''' warnings.warn( 'The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ImageGPTImageProcessor instead.' , __snake_case , ) super().__init__(*__snake_case , **__snake_case )
218
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCamelCase_( _snake_case : Dict , _snake_case : Optional[int] , _snake_case : str ): """simple docstring""" if openai_config_file == "": __a =OpenAIGPTConfig() else: __a =OpenAIGPTConfig.from_json_file(_snake_case ) __a =OpenAIGPTModel(_snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(_snake_case , _snake_case , _snake_case ) # Save pytorch-model __a =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __a =pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , _snake_case ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(_snake_case , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--openai_checkpoint_folder_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--openai_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained OpenAI model. \n" "This specifies the model architecture." ), ) _lowerCAmelCase : int = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
218
1
from __future__ import annotations from scipy.special import comb # type: ignore class _lowerCAmelCase: """simple docstring""" def __init__( self , _lowerCamelCase ): UpperCamelCase_: int = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. UpperCamelCase_: Dict = len(lowercase_ ) - 1 def _a ( self , _lowerCamelCase ): assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCamelCase_: list[float] = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , lowercase_ ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(lowercase_ ) , 5 ) == 1 return output_values def _a ( self , _lowerCamelCase ): assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCamelCase_: List[str] = self.basis_function(lowercase_ ) UpperCamelCase_: Tuple = 0.0 UpperCamelCase_: Dict = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def _a ( self , _lowerCamelCase = 0.0_1 ): from matplotlib import pyplot as plt # type: ignore UpperCamelCase_: list[float] = [] # x coordinates of points to plot UpperCamelCase_: list[float] = [] # y coordinates of points to plot UpperCamelCase_: Tuple = 0.0 while t <= 1: UpperCamelCase_: Any = self.bezier_curve_function(lowercase_ ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size UpperCamelCase_: Any = [i[0] for i in self.list_of_points] UpperCamelCase_: List[str] = [i[1] for i in self.list_of_points] plt.plot( lowercase_ , lowercase_ , color='blue' , label='Curve of Degree ' + str(self.degree ) , ) plt.scatter(lowercase_ , lowercase_ , color='red' , label='Control Points' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
371
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" a : Optional[int] =LongformerTokenizer a : Optional[int] =True a : Tuple =LongformerTokenizerFast a : Dict =True def _a ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase_: int = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] UpperCamelCase_: Optional[Any] = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) UpperCamelCase_: Any = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] UpperCamelCase_: Tuple = {'unk_token': '<unk>'} UpperCamelCase_: str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCamelCase_: Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_lowerCamelCase ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_lowerCamelCase ) ) def _a ( self , **_lowerCamelCase ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def _a ( self , **_lowerCamelCase ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def _a ( self , _lowerCamelCase ): UpperCamelCase_: Union[str, Any] = 'lower newer' UpperCamelCase_: Optional[Any] = 'lower newer' return input_text, output_text def _a ( self ): UpperCamelCase_: int = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase_: Any = 'lower newer' UpperCamelCase_: Any = ['l', 'o', 'w', 'er', '\u0120', 'n', 'e', 'w', 'er'] UpperCamelCase_: str = tokenizer.tokenize(_lowerCamelCase ) # , add_prefix_space=True) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) UpperCamelCase_: Dict = tokens + [tokenizer.unk_token] UpperCamelCase_: int = [0, 1, 2, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , _lowerCamelCase ) def _a ( self ): UpperCamelCase_: Tuple = self.get_tokenizer() self.assertListEqual(tokenizer.encode('Hello world!' , add_special_tokens=_lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 2] ) self.assertListEqual( tokenizer.encode('Hello world! cécé herlolip 418' , add_special_tokens=_lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2] , ) @slow def _a ( self ): UpperCamelCase_: int = self.tokenizer_class.from_pretrained('allenai/longformer-base-4096' ) UpperCamelCase_: Dict = tokenizer.encode('sequence builders' , add_special_tokens=_lowerCamelCase ) UpperCamelCase_: Any = tokenizer.encode('multi-sequence build' , add_special_tokens=_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = tokenizer.encode( 'sequence builders' , add_special_tokens=_lowerCamelCase , add_prefix_space=_lowerCamelCase ) UpperCamelCase_: Any = tokenizer.encode( 'sequence builders' , 'multi-sequence build' , add_special_tokens=_lowerCamelCase , add_prefix_space=_lowerCamelCase ) UpperCamelCase_: int = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase ) UpperCamelCase_: Tuple = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase , _lowerCamelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def _a ( self ): UpperCamelCase_: Optional[int] = self.get_tokenizer() UpperCamelCase_: Optional[int] = 'Encode this sequence.' UpperCamelCase_: List[Any] = tokenizer.byte_encoder[' '.encode('utf-8' )[0]] # Testing encoder arguments UpperCamelCase_: Dict = tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase , add_prefix_space=_lowerCamelCase ) UpperCamelCase_: List[str] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(_lowerCamelCase , _lowerCamelCase ) UpperCamelCase_: List[Any] = tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase , add_prefix_space=_lowerCamelCase ) UpperCamelCase_: int = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(_lowerCamelCase , _lowerCamelCase ) tokenizer.add_special_tokens({'bos_token': '<s>'} ) UpperCamelCase_: Optional[Any] = tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase ) UpperCamelCase_: Any = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(_lowerCamelCase , _lowerCamelCase ) # Testing spaces after special tokens UpperCamelCase_: List[Any] = '<mask>' tokenizer.add_special_tokens( {'mask_token': AddedToken(_lowerCamelCase , lstrip=_lowerCamelCase , rstrip=_lowerCamelCase )} ) # mask token has a left space UpperCamelCase_: List[Any] = tokenizer.convert_tokens_to_ids(_lowerCamelCase ) UpperCamelCase_: Dict = 'Encode <mask> sequence' UpperCamelCase_: Dict = 'Encode <mask>sequence' UpperCamelCase_: Any = tokenizer.encode(_lowerCamelCase ) UpperCamelCase_: Optional[Any] = encoded.index(_lowerCamelCase ) UpperCamelCase_: Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(_lowerCamelCase , _lowerCamelCase ) UpperCamelCase_: List[str] = tokenizer.encode(_lowerCamelCase ) UpperCamelCase_: List[Any] = encoded.index(_lowerCamelCase ) UpperCamelCase_: Optional[int] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(_lowerCamelCase , _lowerCamelCase ) def _a ( self ): pass def _a ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCamelCase_: Any = self.rust_tokenizer_class.from_pretrained(_lowerCamelCase , **_lowerCamelCase ) UpperCamelCase_: Tuple = self.tokenizer_class.from_pretrained(_lowerCamelCase , **_lowerCamelCase ) UpperCamelCase_: List[Any] = 'A, <mask> AllenNLP sentence.' UpperCamelCase_: int = tokenizer_r.encode_plus(_lowerCamelCase , add_special_tokens=_lowerCamelCase , return_token_type_ids=_lowerCamelCase ) UpperCamelCase_: Any = tokenizer_p.encode_plus(_lowerCamelCase , add_special_tokens=_lowerCamelCase , return_token_type_ids=_lowerCamelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids'] ) , sum(tokens_p['token_type_ids'] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask'] ) / len(tokens_r['attention_mask'] ) , sum(tokens_p['attention_mask'] ) / len(tokens_p['attention_mask'] ) , ) UpperCamelCase_: List[str] = tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids'] ) UpperCamelCase_: str = tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids'] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual(tokens_r['input_ids'] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual( _lowerCamelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) self.assertSequenceEqual( _lowerCamelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) def _a ( self ): for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): UpperCamelCase_: Optional[Any] = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) UpperCamelCase_: Optional[int] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['add_prefix_space'] , _lowerCamelCase ) self.assertEqual(post_processor_state['add_prefix_space'] , _lowerCamelCase ) self.assertEqual(post_processor_state['trim_offsets'] , _lowerCamelCase ) def _a ( self ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCamelCase_: Optional[int] = 'hello' # `hello` is a token in the vocabulary of `pretrained_name` UpperCamelCase_: Union[str, Any] = f'''{text_of_1_token} {text_of_1_token}''' UpperCamelCase_: Optional[int] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: str = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCamelCase ) + 1, len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: Optional[int] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: str = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCamelCase ) + 1, len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: Optional[Any] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Dict = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCamelCase ), len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: List[str] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Optional[Any] = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCamelCase ), len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: Optional[int] = f''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) UpperCamelCase_: Optional[int] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Tuple = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCamelCase ) + 1, 1 + len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: Dict = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Dict = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCamelCase ), 1 + len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , ) UpperCamelCase_: Union[str, Any] = self.rust_tokenizer_class.from_pretrained( _lowerCamelCase , use_fast=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = tokenizer_r(_lowerCamelCase , return_offsets_mapping=_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCamelCase ), 1 + len(_lowerCamelCase ) + 1 + len(_lowerCamelCase )) , )
292
0
"""simple docstring""" import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def A_ ( _lowercase, _lowercase ): '''simple docstring''' assert isinstance(lowercase_, lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""", [False, True] ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :str = tmp_path / """cache""" snake_case_ :Optional[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): snake_case_ :Union[str, Any] = ParquetDatasetReader(lowercase_, cache_dir=lowercase_, keep_in_memory=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) @pytest.mark.parametrize( """features""", [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ], ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Any = tmp_path / """cache""" snake_case_ :List[str] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} snake_case_ :List[Any] = features.copy() if features else default_expected_features snake_case_ :Optional[Any] = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) snake_case_ :Tuple = ParquetDatasetReader(lowercase_, features=lowercase_, cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) @pytest.mark.parametrize("""split""", [None, NamedSplit("""train""" ), """train""", """test"""] ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :List[str] = tmp_path / """cache""" snake_case_ :Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} snake_case_ :Dict = ParquetDatasetReader(lowercase_, cache_dir=lowercase_, split=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("""path_type""", [str, list] ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' if issubclass(lowercase_, lowercase_ ): snake_case_ :Tuple = parquet_path elif issubclass(lowercase_, lowercase_ ): snake_case_ :int = [parquet_path] snake_case_ :Optional[Any] = tmp_path / """cache""" snake_case_ :str = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} snake_case_ :Optional[Any] = ParquetDatasetReader(lowercase_, cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) def A_ ( _lowercase, _lowercase, _lowercase=("train",) ): '''simple docstring''' assert isinstance(lowercase_, lowercase_ ) for split in splits: snake_case_ :Tuple = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""", [False, True] ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Dict = tmp_path / """cache""" snake_case_ :int = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): snake_case_ :List[Any] = ParquetDatasetReader( {"""train""": parquet_path}, cache_dir=lowercase_, keep_in_memory=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_ ) @pytest.mark.parametrize( """features""", [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ], ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Dict = tmp_path / """cache""" snake_case_ :List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} snake_case_ :Optional[Any] = features.copy() if features else default_expected_features snake_case_ :List[Any] = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) snake_case_ :List[Any] = ParquetDatasetReader({"""train""": parquet_path}, features=lowercase_, cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_ ) @pytest.mark.parametrize("""split""", [None, NamedSplit("""train""" ), """train""", """test"""] ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' if split: snake_case_ :int = {split: parquet_path} else: snake_case_ :List[str] = """train""" snake_case_ :Dict = {"""train""": parquet_path, """test""": parquet_path} snake_case_ :Optional[Any] = tmp_path / """cache""" snake_case_ :List[str] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} snake_case_ :Dict = ParquetDatasetReader(lowercase_, cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_, splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def A_ ( _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Tuple = ParquetDatasetWriter(lowercase_, tmp_path / """foo.parquet""" ) assert writer.write() > 0 snake_case_ :int = pq.ParquetFile(tmp_path / """foo.parquet""" ) snake_case_ :Optional[Any] = pf.read() assert dataset.data.table == output_table def A_ ( _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Union[str, Any] = str(shared_datadir / """test_image_rgb.jpg""" ) snake_case_ :Any = {"""image""": [image_path]} snake_case_ :Optional[Any] = Features({"""image""": Image()} ) snake_case_ :str = Dataset.from_dict(lowercase_, features=lowercase_ ) snake_case_ :str = ParquetDatasetWriter(lowercase_, tmp_path / """foo.parquet""" ) assert writer.write() > 0 snake_case_ :int = Dataset.from_parquet(str(tmp_path / """foo.parquet""" ) ) assert dataset.features == reloaded_dataset.features snake_case_ :List[Any] = ParquetDatasetReader(str(tmp_path / """foo.parquet""" ), streaming=lowercase_ ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( """feature, expected""", [ (Features({"""foo""": Value("""int32""" )} ), None), (Features({"""image""": Image(), """foo""": Value("""int32""" )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({"""nested""": Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ], ) def A_ ( _lowercase, _lowercase ): '''simple docstring''' assert get_writer_batch_size(lowercase_ ) == expected
66
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( """files""" , [ ["""full:README.md""", """dataset_infos.json"""], ["""empty:README.md""", """dataset_infos.json"""], ["""dataset_infos.json"""], ["""full:README.md"""], ] , ) def UpperCamelCase (lowercase_: List[str] , lowercase_: Optional[int] ) -> List[str]: A__ : Tuple = tmp_path_factory.mktemp("""dset_infos_dir""" ) if "full:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""---\ndataset_info:\n dataset_size: 42\n---""" ) if "empty:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""""" ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / """dataset_infos.json""" , """w""" ) as f: f.write("""{\"default\": {\"dataset_size\": 42}}""" ) A__ : List[Any] = DatasetInfosDict.from_directory(lowercase_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( """dataset_info""" , [ DatasetInfo(), DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ), ] , ) def UpperCamelCase (lowercase_: str , lowercase_: DatasetInfo ) -> List[Any]: A__ : Union[str, Any] = str(lowercase_ ) dataset_info.write_to_directory(lowercase_ ) A__ : List[Any] = DatasetInfo.from_directory(lowercase_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(lowercase_ , """dataset_info.json""" ) ) def UpperCamelCase () -> List[Any]: A__ : Union[str, Any] = DatasetInfo( description="""foo""" , citation="""bar""" , homepage="""https://foo.bar""" , license="""CC0""" , features=Features({"""a""": Value("""int32""" )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train""", """num_examples""": 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) A__ : Dict = dataset_info._to_yaml_dict() assert sorted(lowercase_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) A__ : Union[str, Any] = yaml.safe_dump(lowercase_ ) A__ : List[Any] = yaml.safe_load(lowercase_ ) assert dataset_info_yaml_dict == reloaded def UpperCamelCase () -> List[str]: A__ : Optional[int] = DatasetInfo() A__ : List[Any] = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( """dataset_infos_dict""" , [ DatasetInfosDict(), DatasetInfosDict({"""default""": DatasetInfo()} ), DatasetInfosDict({"""my_config_name""": DatasetInfo()} ), DatasetInfosDict( { """default""": DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ) } ), DatasetInfosDict( { """v1""": DatasetInfo(dataset_size=42 ), """v2""": DatasetInfo(dataset_size=1337 ), } ), ] , ) def UpperCamelCase (lowercase_: Tuple , lowercase_: DatasetInfosDict ) -> Optional[Any]: A__ : List[Any] = str(lowercase_ ) dataset_infos_dict.write_to_directory(lowercase_ ) A__ : Dict = DatasetInfosDict.from_directory(lowercase_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): A__ : Optional[int] = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml A__ : List[str] = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(lowercase_ , """README.md""" ) )
192
0
'''simple docstring''' from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
13
'''simple docstring''' from __future__ import annotations def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> None: """simple docstring""" __snake_case : int = len(_lowerCamelCase ) # If row is equal to the size of the board it means there are a queen in each row in # the current board (possible_board) if row == n: # We convert the variable possible_board that looks like this: [1, 3, 0, 2] to # this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . '] boards.append([""". """ * i + """Q """ + """. """ * (n - 1 - i) for i in possible_board] ) return # We iterate each column in the row to find all possible results in each row for col in range(_lowerCamelCase ): # We apply that we learned previously. First we check that in the current board # (possible_board) there are not other same value because if there is it means # that there are a collision in vertical. Then we apply the two formulas we # learned before: # # 45º: y - x = b or 45: row - col = b # 135º: y + x = b or row + col = b. # # And we verify if the results of this two formulas not exist in their variables # respectively. (diagonal_right_collisions, diagonal_left_collisions) # # If any or these are True it means there is a collision so we continue to the # next value in the for loop. if ( col in possible_board or row - col in diagonal_right_collisions or row + col in diagonal_left_collisions ): continue # If it is False we call dfs function again and we update the inputs depth_first_search( [*possible_board, col] , [*diagonal_right_collisions, row - col] , [*diagonal_left_collisions, row + col] , _lowerCamelCase , _lowerCamelCase , ) def _a ( _lowerCamelCase ) -> None: """simple docstring""" __snake_case : list[list[str]] = [] depth_first_search([] , [] , [] , _lowerCamelCase , _lowerCamelCase ) # Print all the boards for board in boards: for column in board: print(_lowerCamelCase ) print("""""" ) print(len(_lowerCamelCase ) , """solutions were found.""" ) if __name__ == "__main__": import doctest doctest.testmod() n_queens_solution(4)
13
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class _UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self :Tuple ): A = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) A = get_activation("gelu" ) self.assertTrue(torch.allclose(gelu_python(__UpperCamelCase ) , torch_builtin(__UpperCamelCase ) ) ) self.assertFalse(torch.allclose(gelu_python(__UpperCamelCase ) , gelu_new(__UpperCamelCase ) ) ) def lowerCamelCase ( self :Union[str, Any] ): A = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) A = get_activation("gelu" ) A = get_activation("gelu_10" ) A = torch_builtin(__UpperCamelCase ) A = geluaa(__UpperCamelCase ) A = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(__UpperCamelCase ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCamelCase ( self :Dict ): get_activation("gelu" ) get_activation("gelu_10" ) get_activation("gelu_fast" ) get_activation("gelu_new" ) get_activation("gelu_python" ) get_activation("gelu_pytorch_tanh" ) get_activation("linear" ) get_activation("mish" ) get_activation("quick_gelu" ) get_activation("relu" ) get_activation("sigmoid" ) get_activation("silu" ) get_activation("swish" ) get_activation("tanh" ) with self.assertRaises(__UpperCamelCase ): get_activation("bogus" ) with self.assertRaises(__UpperCamelCase ): get_activation(__UpperCamelCase ) def lowerCamelCase ( self :str ): A = get_activation("gelu" ) A = 1 A = get_activation("gelu" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(__UpperCamelCase ): A = acta.a
292
"""simple docstring""" _snake_case : Optional[int] = [ 'DownloadConfig', 'DownloadManager', 'DownloadMode', 'StreamingDownloadManager', ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
292
1
from __future__ import annotations from typing import TypedDict class SCREAMING_SNAKE_CASE_ ( snake_case_ ): __magic_name__: str __magic_name__: int def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): raise TypeError('The parameter s type must be str.' ) return [s[i:] + s[:i] for i in range(len(__a ) )] def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): raise TypeError('The parameter s type must be str.' ) if not s: raise ValueError('The parameter s must not be empty.' ) snake_case_ : Optional[Any] = all_rotations(__a ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation snake_case_ : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__a ), } return response def SCREAMING_SNAKE_CASE__ ( __a , __a ): if not isinstance(__a , __a ): raise TypeError('The parameter bwt_string type must be str.' ) if not bwt_string: raise ValueError('The parameter bwt_string must not be empty.' ) try: snake_case_ : str = int(__a ) except ValueError: raise TypeError( 'The parameter idx_original_string type must be int or passive' ' of cast to int.' ) if idx_original_string < 0: raise ValueError('The parameter idx_original_string must not be lower than 0.' ) if idx_original_string >= len(__a ): raise ValueError( 'The parameter idx_original_string must be lower than' ' len(bwt_string).' ) snake_case_ : Union[str, Any] = [''] * len(__a ) for _ in range(len(__a ) ): for i in range(len(__a ) ): snake_case_ : Dict = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": _SCREAMING_SNAKE_CASE = """Provide a string that I will generate its BWT transform: """ _SCREAMING_SNAKE_CASE = input(entry_msg).strip() _SCREAMING_SNAKE_CASE = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result['bwt_string']}\'''' ) _SCREAMING_SNAKE_CASE = reverse_bwt(result["""bwt_string"""], result["""idx_original_string"""]) print( F'''Reversing Burrows Wheeler transform for entry \'{result['bwt_string']}\' ''' F'''we get original string \'{original_string}\'''' )
88
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def SCREAMING_SNAKE_CASE__ ( __a , __a ): snake_case_ : List[Any] = k_size // 2 snake_case_ ,snake_case_ : Any = mgrid[0 - center : k_size - center, 0 - center : k_size - center] snake_case_ : Optional[Any] = 1 / (2 * pi * sigma) * exp(-(square(__a ) + square(__a )) / (2 * square(__a )) ) return g def SCREAMING_SNAKE_CASE__ ( __a , __a , __a ): snake_case_ ,snake_case_ : Union[str, Any] = image.shape[0], image.shape[1] # dst image height and width snake_case_ : int = height - k_size + 1 snake_case_ : Optional[int] = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows snake_case_ : Optional[Any] = zeros((dst_height * dst_width, k_size * k_size) ) snake_case_ : Tuple = 0 for i, j in product(range(__a ) , range(__a ) ): snake_case_ : Optional[int] = ravel(image[i : i + k_size, j : j + k_size] ) snake_case_ : str = window row += 1 # turn the kernel into shape(k*k, 1) snake_case_ : List[Any] = gen_gaussian_kernel(__a , __a ) snake_case_ : str = ravel(__a ) # reshape and get the dst image snake_case_ : Optional[int] = dot(__a , __a ).reshape(__a , __a ).astype(__a ) return dst if __name__ == "__main__": # read original image _SCREAMING_SNAKE_CASE = imread(R"""../image_data/lena.jpg""") # turn image in gray scale value _SCREAMING_SNAKE_CASE = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size _SCREAMING_SNAKE_CASE = gaussian_filter(gray, 3, sigma=1) _SCREAMING_SNAKE_CASE = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow("""gaussian filter with 3x3 mask""", gaussianaxa) imshow("""gaussian filter with 5x5 mask""", gaussianaxa) waitKey()
88
1
def __snake_case ( _lowerCAmelCase : List[str] ) -> Optional[Any]: A_ : Dict = [0] * len(_lowerCAmelCase ) A_ : str = [] A_ : Dict = [1] * len(_lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_lowerCAmelCase ) ): if indegree[i] == 0: queue.append(_lowerCAmelCase ) while queue: A_ : Any = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: A_ : Tuple = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(_lowerCAmelCase ) print(max(_lowerCAmelCase ) ) # Adjacency list of Graph _lowerCAmelCase : List[str] = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
300
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore _lowerCAmelCase : Optional[int] = ''' Human: <<task>> Assistant: ''' _lowerCAmelCase : int = '''huggingface-tools/default-prompts''' _lowerCAmelCase : Any = {'''chat''': '''chat_prompt_template.txt''', '''run''': '''run_prompt_template.txt'''} def __snake_case ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict="run" ) -> List[Any]: if prompt_or_repo_id is None: A_ : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , _lowerCAmelCase ) is not None: return prompt_or_repo_id A_ : Optional[Any] = cached_file( _lowerCAmelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(_lowerCAmelCase , "r" , encoding="utf-8" ) as f: return f.read()
300
1
import sys def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[Any] = len(lowerCamelCase_ ) _lowercase : int = [[0 for x in range(lowerCamelCase_ )] for x in range(lowerCamelCase_ )] _lowercase : Dict = [[0 for x in range(lowerCamelCase_ )] for x in range(lowerCamelCase_ )] for chain_length in range(2 , lowerCamelCase_ ): for a in range(1 , n - chain_length + 1 ): _lowercase : str = a + chain_length - 1 _lowercase : Tuple = sys.maxsize for c in range(lowerCamelCase_ , lowerCamelCase_ ): _lowercase : Dict = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: _lowercase : Any = cost _lowercase : Dict = c return matrix, sol def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: if i == j: print('A' + str(lowerCamelCase_ ) , end=' ' ) else: print('(' , end=' ' ) print_optiomal_solution(lowerCamelCase_ , lowerCamelCase_ , optimal_solution[i][j] ) print_optiomal_solution(lowerCamelCase_ , optimal_solution[i][j] + 1 , lowerCamelCase_ ) print(')' , end=' ' ) def UpperCamelCase_( ) -> Optional[Any]: _lowercase : Optional[int] = [30, 35, 15, 5, 10, 20, 25] _lowercase : List[str] = len(lowerCamelCase_ ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 _lowercase , _lowercase : Tuple = matrix_chain_order(lowerCamelCase_ ) print('No. of Operation required: ' + str(matrix[1][n - 1] ) ) print_optiomal_solution(lowerCamelCase_ , 1 , n - 1 ) if __name__ == "__main__": main()
84
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE : str = logging.getLogger() def UpperCamelCase_( ) -> Any: _lowercase : int = argparse.ArgumentParser() parser.add_argument('-f' ) _lowercase : Optional[Any] = parser.parse_args() return args.f class _lowerCamelCase( _a ): def UpperCamelCase ( self) -> None: """simple docstring""" _lowercase : List[Any] = logging.StreamHandler(sys.stdout) logger.addHandler(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : str = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0, 'run_glue_deebert.py') with patch.object(lowerCamelCase, 'argv', lowerCamelCase): _lowercase : Optional[Any] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase, 0.6_6_6) @slow @require_torch_non_multi_gpu def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Union[str, Any] = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(lowerCamelCase) _lowercase : Union[str, Any] = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(lowerCamelCase) _lowercase : Union[str, Any] = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(lowerCamelCase)
84
1
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys __a = '''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
337
import importlib.metadata import operator import re import sys from typing import Optional from packaging import version __a = { '''<''': operator.lt, '''<=''': operator.le, '''==''': operator.eq, '''!=''': operator.ne, '''>=''': operator.ge, '''>''': operator.gt, } def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->Optional[int]: """simple docstring""" if got_ver is None or want_ver is None: raise ValueError( f"""Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider""" f""" reinstalling {pkg}.""" ) if not ops[op](version.parse(_UpperCamelCase ), version.parse(_UpperCamelCase ) ): raise ImportError( f"""{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}""" ) def __lowercase ( _UpperCamelCase, _UpperCamelCase = None ) ->None: """simple docstring""" lowercase : List[Any] = f"""\n{hint}""" if hint is not None else '''''' # non-versioned check if re.match(R'''^[\w_\-\d]+$''', _UpperCamelCase ): lowercase , lowercase , lowercase : Optional[Any] = requirement, None, None else: lowercase : List[Any] = re.findall(R'''^([^!=<>\s]+)([\s!=<>]{1,2}.+)''', _UpperCamelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but''' f""" got {requirement}""" ) lowercase , lowercase : str = match[0] lowercase : Tuple = want_full.split(''',''' ) # there could be multiple requirements lowercase : List[Any] = {} for w in want_range: lowercase : str = re.findall(R'''^([\s!=<>]{1,2})(.+)''', _UpperCamelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,''' f""" but got {requirement}""" ) lowercase , lowercase : Optional[int] = match[0] lowercase : Dict = want_ver if op not in ops: raise ValueError(f"""{requirement}: need one of {list(ops.keys() )}, but got {op}""" ) # special case if pkg == "python": lowercase : int = '''.'''.join([str(_UpperCamelCase ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(_UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) return # check if any version is installed try: lowercase : List[str] = importlib.metadata.version(_UpperCamelCase ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( f"""The '{requirement}' distribution was not found and is required by this application. {hint}""" ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(_UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) def __lowercase ( _UpperCamelCase ) ->int: """simple docstring""" lowercase : Optional[int] = '''Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main''' return require_version(_UpperCamelCase, _UpperCamelCase )
337
1
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCamelCase__ = "hf-internal-testing/tiny-random-bert" lowerCamelCase__ = os.path.join(TRANSFORMERS_CACHE, "models--hf-internal-testing--tiny-random-bert") lowerCamelCase__ = "9b8c223d42b2188cb49d29af482996f9d0f3e5a6" class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Any: _UpperCamelCase : Any = cached_file(__a , __a ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(__a ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(__a , __a ) ) ) with open(os.path.join(__a , "refs" , "main" ) ) as f: _UpperCamelCase : Dict = f.read() self.assertEqual(__a , os.path.join(__a , "snapshots" , __a , __a ) ) self.assertTrue(os.path.isfile(__a ) ) # File is cached at the same place the second time. _UpperCamelCase : Tuple = cached_file(__a , __a ) self.assertEqual(__a , __a ) # Using a specific revision to test the full commit hash. _UpperCamelCase : Any = cached_file(__a , __a , revision="9b8c223" ) self.assertEqual(__a , os.path.join(__a , "snapshots" , __a , __a ) ) def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[str]: with self.assertRaisesRegex(__a , "is not a valid model identifier" ): _UpperCamelCase : Tuple = cached_file("tiny-random-bert" , __a ) with self.assertRaisesRegex(__a , "is not a valid git identifier" ): _UpperCamelCase : int = cached_file(__a , __a , revision="aaaa" ) with self.assertRaisesRegex(__a , "does not appear to have a file named" ): _UpperCamelCase : Tuple = cached_file(__a , "conf" ) def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: with self.assertRaisesRegex(__a , "does not appear to have a file named" ): _UpperCamelCase : List[Any] = cached_file(__a , "conf" ) with open(os.path.join(__a , "refs" , "main" ) ) as f: _UpperCamelCase : Optional[Any] = f.read() self.assertTrue(os.path.isfile(os.path.join(__a , ".no_exist" , __a , "conf" ) ) ) _UpperCamelCase : List[Any] = cached_file(__a , "conf" , _raise_exceptions_for_missing_entries=__a ) self.assertIsNone(__a ) _UpperCamelCase : int = cached_file(__a , "conf" , local_files_only=__a , _raise_exceptions_for_missing_entries=__a ) self.assertIsNone(__a ) _UpperCamelCase : Any = mock.Mock() _UpperCamelCase : Dict = 500 _UpperCamelCase : str = {} _UpperCamelCase : Optional[Any] = HTTPError _UpperCamelCase : Optional[Any] = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=__a ) as mock_head: _UpperCamelCase : Union[str, Any] = cached_file(__a , "conf" , _raise_exceptions_for_connection_errors=__a ) self.assertIsNone(__a ) # This check we did call the fake head request mock_head.assert_called() def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[int]: self.assertTrue(has_file("hf-internal-testing/tiny-bert-pt-only" , __a ) ) self.assertFalse(has_file("hf-internal-testing/tiny-bert-pt-only" , __a ) ) self.assertFalse(has_file("hf-internal-testing/tiny-bert-pt-only" , __a ) ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[Any]: # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo("bert-base-cased" , "ahah.txt" ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(__a , "is not a valid model identifier" ): get_file_from_repo("bert-base-case" , __a ) # The function raises if the revision does not exist. with self.assertRaisesRegex(__a , "is not a valid git identifier" ): get_file_from_repo("bert-base-cased" , __a , revision="ahaha" ) _UpperCamelCase : Tuple = get_file_from_repo("bert-base-cased" , __a ) # The name is the cached name which is not very easy to test, so instead we load the content. _UpperCamelCase : List[Any] = json.loads(open(__a , "r" ).read() ) self.assertEqual(config["hidden_size"] , 768 ) def __SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: with tempfile.TemporaryDirectory() as tmp_dir: _UpperCamelCase : Union[str, Any] = Path(__a ) / "a.txt" filename.touch() self.assertEqual(get_file_from_repo(__a , "a.txt" ) , str(__a ) ) self.assertIsNone(get_file_from_repo(__a , "b.txt" ) )
310
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Dict , __a : List[Any] , __a : str=13 , __a : Any=30 , __a : List[str]=2 , __a : Dict=3 , __a : Union[str, Any]=True , __a : Dict=True , __a : List[str]=32 , __a : Tuple=5 , __a : str=4 , __a : List[str]=37 , __a : Tuple="gelu" , __a : str=0.1 , __a : Optional[int]=0.1 , __a : Union[str, Any]=10 , __a : Optional[Any]=0.02 , __a : List[Any]=None , __a : str=2 , ) -> int: _UpperCamelCase : Tuple = parent _UpperCamelCase : str = batch_size _UpperCamelCase : Tuple = image_size _UpperCamelCase : List[str] = patch_size _UpperCamelCase : Dict = num_channels _UpperCamelCase : List[str] = is_training _UpperCamelCase : Any = use_labels _UpperCamelCase : int = hidden_size _UpperCamelCase : List[Any] = num_hidden_layers _UpperCamelCase : Union[str, Any] = num_attention_heads _UpperCamelCase : Optional[int] = intermediate_size _UpperCamelCase : Any = hidden_act _UpperCamelCase : Dict = hidden_dropout_prob _UpperCamelCase : Dict = attention_probs_dropout_prob _UpperCamelCase : Optional[int] = type_sequence_label_size _UpperCamelCase : int = initializer_range _UpperCamelCase : Optional[int] = scope _UpperCamelCase : Any = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCamelCase : Optional[int] = (image_size // patch_size) ** 2 _UpperCamelCase : Optional[int] = num_patches + 1 def __SCREAMING_SNAKE_CASE ( self : int ) -> Optional[Any]: _UpperCamelCase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCamelCase : Union[str, Any] = None if self.use_labels: _UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCamelCase : Any = self.get_config() return config, pixel_values, labels def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[str]: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __SCREAMING_SNAKE_CASE ( self : Tuple , __a : Optional[int] , __a : Union[str, Any] , __a : Tuple ) -> Union[str, Any]: _UpperCamelCase : Optional[Any] = ViTModel(config=__a ) model.to(__a ) model.eval() _UpperCamelCase : Tuple = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __SCREAMING_SNAKE_CASE ( self : Dict , __a : str , __a : Optional[int] , __a : int ) -> Optional[int]: _UpperCamelCase : Tuple = ViTForMaskedImageModeling(config=__a ) model.to(__a ) model.eval() _UpperCamelCase : Any = model(__a ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _UpperCamelCase : Union[str, Any] = 1 _UpperCamelCase : Union[str, Any] = ViTForMaskedImageModeling(__a ) model.to(__a ) model.eval() _UpperCamelCase : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _UpperCamelCase : Dict = model(__a ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __SCREAMING_SNAKE_CASE ( self : List[Any] , __a : Tuple , __a : int , __a : Dict ) -> int: _UpperCamelCase : Any = self.type_sequence_label_size _UpperCamelCase : Optional[Any] = ViTForImageClassification(__a ) model.to(__a ) model.eval() _UpperCamelCase : int = model(__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _UpperCamelCase : Tuple = 1 _UpperCamelCase : Union[str, Any] = ViTForImageClassification(__a ) model.to(__a ) model.eval() _UpperCamelCase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _UpperCamelCase : List[Any] = model(__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __SCREAMING_SNAKE_CASE ( self : str ) -> Tuple: _UpperCamelCase : Dict = self.prepare_config_and_inputs() ( ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ) : Union[str, Any] = config_and_inputs _UpperCamelCase : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :Optional[Any] = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ :Any = ( {"feature-extraction": ViTModel, "image-classification": ViTForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ :str = True SCREAMING_SNAKE_CASE__ :List[Any] = False SCREAMING_SNAKE_CASE__ :int = False SCREAMING_SNAKE_CASE__ :int = False def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: _UpperCamelCase : Dict = ViTModelTester(self ) _UpperCamelCase : Any = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37 ) def __SCREAMING_SNAKE_CASE ( self : str ) -> Optional[Any]: self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def __SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: pass def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Union[str, Any]: _UpperCamelCase, _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : List[Any] = model_class(__a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCamelCase : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear ) ) def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[Any]: _UpperCamelCase, _UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : Any = model_class(__a ) _UpperCamelCase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase : List[str] = [*signature.parameters.keys()] _UpperCamelCase : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , __a ) def __SCREAMING_SNAKE_CASE ( self : Any ) -> int: _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def __SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: _UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__a ) def __SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: _UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__a ) @slow def __SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase : List[str] = ViTModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def lowercase__ ( ) -> str: """simple docstring""" _UpperCamelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[int]: return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def __SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: _UpperCamelCase : List[Any] = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(__a ) _UpperCamelCase : str = self.default_image_processor _UpperCamelCase : List[Any] = prepare_img() _UpperCamelCase : Any = image_processor(images=__a , return_tensors="pt" ).to(__a ) # forward pass with torch.no_grad(): _UpperCamelCase : Dict = model(**__a ) # verify the logits _UpperCamelCase : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __a ) _UpperCamelCase : str = torch.tensor([-0.27_44, 0.82_15, -0.08_36] ).to(__a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4 ) ) @slow def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> str: # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. _UpperCamelCase : List[str] = ViTModel.from_pretrained("facebook/dino-vits8" ).to(__a ) _UpperCamelCase : Union[str, Any] = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) _UpperCamelCase : List[str] = prepare_img() _UpperCamelCase : int = image_processor(images=__a , return_tensors="pt" ) _UpperCamelCase : Any = inputs.pixel_values.to(__a ) # forward pass with torch.no_grad(): _UpperCamelCase : str = model(__a , interpolate_pos_encoding=__a ) # verify the logits _UpperCamelCase : int = torch.Size((1, 3601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , __a ) _UpperCamelCase : int = torch.tensor( [[4.23_40, 4.39_06, -6.66_92], [4.54_63, 1.89_28, -6.72_57], [4.44_29, 0.84_96, -5.85_85]] ).to(__a ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __a , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Any: _UpperCamelCase : Tuple = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) _UpperCamelCase : int = self.default_image_processor _UpperCamelCase : Dict = prepare_img() _UpperCamelCase : Union[str, Any] = image_processor(images=__a , return_tensors="pt" ) _UpperCamelCase : Any = inputs.pixel_values.to(__a ) # forward pass to make sure inference works in fp16 with torch.no_grad(): _UpperCamelCase : int = model(__a )
310
1
def A_ ( A__ ) -> None: a__ : Optional[int] = generate_pascal_triangle(A__ ) for row_idx in range(A__ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=' ' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=' ' ) else: print(triangle[row_idx][col_idx] , end='' ) print() def A_ ( A__ ) -> list[list[int]]: if not isinstance(A__ , A__ ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) a__ : list[list[int]] = [] for current_row_idx in range(A__ ): a__ : Any = populate_current_row(A__ , A__ ) triangle.append(A__ ) return triangle def A_ ( A__ , A__ ) -> list[int]: a__ : Optional[int] = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 a__ , a__ : str = 1, 1 for current_col_idx in range(1 , A__ ): calculate_current_element( A__ , A__ , A__ , A__ ) return current_row def A_ ( A__ , A__ , A__ , A__ , ) -> None: a__ : Dict = triangle[current_row_idx - 1][current_col_idx - 1] a__ : List[str] = triangle[current_row_idx - 1][current_col_idx] a__ : str = above_to_left_elt + above_to_right_elt def A_ ( A__ ) -> list[list[int]]: if not isinstance(A__ , A__ ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) a__ : list[list[int]] = [[1]] for row_index in range(1 , A__ ): a__ : str = [0] + result[-1] + [0] a__ : str = row_index + 1 # Calculate the number of distinct elements in a row a__ : Optional[int] = sum(divmod(A__ , 2 ) ) a__ : Dict = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] a__ : Optional[Any] = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() a__ : Any = row_first_half + row_second_half result.append(A__ ) return result def A_ ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(A__ , A__ ) -> None: a__ : List[str] = F'{func.__name__}({value})' a__ : Tuple = timeit(F'__main__.{call}' , setup='import __main__' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'{call:38} -- {timing:.4f} seconds' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(A__ , A__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
99
import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate UpperCAmelCase_ = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow('', '|', '|'), datarow=DataRow('', '|', '|'), padding=1, with_header_hide=None, ) UpperCAmelCase_ = [] UpperCAmelCase_ = [] UpperCAmelCase_ = {'type': 'section', 'text': {'type': 'plain_text', 'text': 'No failed tests! 🤗', 'emoji': True}} UpperCAmelCase_ = [ { 'type': 'header', 'text': { 'type': 'plain_text', 'text': f"""🤗 Accelerate nightly {os.environ.get("TEST_TYPE", "")} test results""", 'emoji': True, }, } ] UpperCAmelCase_ = 0 for log in Path().glob('*.log'): UpperCAmelCase_ = 0 with open(log, 'r') as f: for line in f: UpperCAmelCase_ = json.loads(line) if line.get('nodeid', '') != "": UpperCAmelCase_ = line['nodeid'] if line.get('duration', None) is not None: UpperCAmelCase_ = f"""{line["duration"]:.4f}""" if line.get('outcome', '') == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split('_')[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) UpperCAmelCase_ = [] log.unlink() UpperCAmelCase_ = '' UpperCAmelCase_ = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += f"*{name[1:]}: {num_failed} failed test*\n" else: message += f"*{name[1:]}: {num_failed} failed tests*\n" UpperCAmelCase_ = [] UpperCAmelCase_ = {} for test in failed_tests: UpperCAmelCase_ = test[0].split('::') UpperCAmelCase_ = data[0].split('/')[-1] if data[0] not in filesafailed: UpperCAmelCase_ = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) UpperCAmelCase_ = [test[0] for test in failed_table] UpperCAmelCase_ = list(set(files)) # Count number of instances in failed_tests UpperCAmelCase_ = [] for file in individual_files: table.append([file, len(filesafailed[file])]) UpperCAmelCase_ = tabulate( table, headers=['Test Location', 'Num Failed'], tablefmt=hf_table_format, stralign='right', ) message += f"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3_000: UpperCAmelCase_ = 'Too many failed tests, please see the full report in the Action results.' UpperCAmelCase_ = len(err) + 10 UpperCAmelCase_ = message[: 3_000 - offset] + f"""\n...\n```\n{err}""" print(f"""### {message}""") else: UpperCAmelCase_ = 'No failed tests! 🤗' print(f"""## {message}""") payload.append(no_error_payload) if os.environ.get('TEST_TYPE', '') != "": from slack_sdk import WebClient UpperCAmelCase_ = WebClient(token=os.environ['SLACK_API_TOKEN']) if message != "No failed tests! 🤗": UpperCAmelCase_ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': message, }, } payload.append(md_report) UpperCAmelCase_ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': '*For more details:*', }, 'accessory': { 'type': 'button', 'text': { 'type': 'plain_text', 'text': 'Check Action results', 'emoji': True, }, 'url': f"""https://github.com/{os.environ["GITHUB_REPOSITORY"]}/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } payload.append(action_button) UpperCAmelCase_ = { 'type': 'context', 'elements': [ { 'type': 'plain_text', 'text': f"""Nightly {os.environ.get("TEST_TYPE")} test results for {date.today()}""", } ], } payload.append(date_report) UpperCAmelCase_ = client.chat_postMessage(channel='#accelerate-ci-daily', text=message, blocks=payload) UpperCAmelCase_ = response.data['ts'] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name UpperCAmelCase_ = '' for i, row in enumerate(test_failures): if row[0] != test_class: UpperCAmelCase_ = row[0] else: UpperCAmelCase_ = '' UpperCAmelCase_ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': f"""Test location: {test_location}\n```\n{tabulate(test_failures, headers=["Class", "Test"], tablefmt=hf_table_format, stralign="right")}\n```""", }, } client.chat_postMessage( channel='#accelerate-ci-daily', thread_ts=ts, blocks=[payload], )
12
0
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def lowerCamelCase_ ( lowerCAmelCase: Optional[Any] )-> List[str]: if isinstance(lowerCAmelCase , collections.abc.Iterable ): return x return (x, x) @require_flax class _lowerCAmelCase : '''simple docstring''' def UpperCamelCase_ ( self : Dict , UpperCamelCase : str , UpperCamelCase : int ): '''simple docstring''' pass def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self : str ): '''simple docstring''' pass def UpperCamelCase_ ( self : str , UpperCamelCase : np.ndarray , UpperCamelCase : np.ndarray , UpperCamelCase : float ): '''simple docstring''' _snake_case : List[Any] = np.abs((a - b) ).max() self.assertLessEqual(UpperCamelCase , UpperCamelCase , f"""Difference between torch and flax is {diff} (>= {tol}).""" ) def UpperCamelCase_ ( self : List[Any] , UpperCamelCase : Any , UpperCamelCase : Tuple , UpperCamelCase : Any , UpperCamelCase : List[str] , UpperCamelCase : Union[str, Any]=None , **UpperCamelCase : Tuple ): '''simple docstring''' _snake_case : Dict = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase , UpperCamelCase ) _snake_case : Optional[int] = FlaxVisionTextDualEncoderModel(UpperCamelCase ) _snake_case : int = model(input_ids=UpperCamelCase , pixel_values=UpperCamelCase , attention_mask=UpperCamelCase ) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], config.projection_dim) ) def UpperCamelCase_ ( self : Dict , UpperCamelCase : List[str] , UpperCamelCase : Any , UpperCamelCase : int , UpperCamelCase : List[Any] , UpperCamelCase : List[Any]=None , **UpperCamelCase : str ): '''simple docstring''' _snake_case , _snake_case : Optional[int] = self.get_vision_text_model(UpperCamelCase , UpperCamelCase ) _snake_case : str = {'vision_model': vision_model, 'text_model': text_model} _snake_case : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase ) _snake_case : Tuple = model(input_ids=UpperCamelCase , pixel_values=UpperCamelCase , attention_mask=UpperCamelCase ) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], model.config.projection_dim) ) def UpperCamelCase_ ( self : str , UpperCamelCase : Optional[int] , UpperCamelCase : Tuple , UpperCamelCase : str , UpperCamelCase : int , UpperCamelCase : Optional[int]=None , **UpperCamelCase : Optional[int] ): '''simple docstring''' _snake_case , _snake_case : Union[str, Any] = self.get_vision_text_model(UpperCamelCase , UpperCamelCase ) _snake_case : List[str] = {'vision_model': vision_model, 'text_model': text_model} _snake_case : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase ) _snake_case : int = model(input_ids=UpperCamelCase , pixel_values=UpperCamelCase , attention_mask=UpperCamelCase ) _snake_case : int = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase ) _snake_case : int = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase ) _snake_case : int = model(input_ids=UpperCamelCase , pixel_values=UpperCamelCase , attention_mask=UpperCamelCase ) _snake_case : int = after_output[0] _snake_case : Union[str, Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase , 1e-3 ) def UpperCamelCase_ ( self : int , UpperCamelCase : Tuple , UpperCamelCase : Optional[Any] , UpperCamelCase : Optional[int] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any]=None , **UpperCamelCase : List[str] ): '''simple docstring''' _snake_case , _snake_case : Optional[Any] = self.get_vision_text_model(UpperCamelCase , UpperCamelCase ) _snake_case : str = {'vision_model': vision_model, 'text_model': text_model} _snake_case : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase ) _snake_case : int = model( input_ids=UpperCamelCase , pixel_values=UpperCamelCase , attention_mask=UpperCamelCase , output_attentions=UpperCamelCase ) _snake_case : Tuple = output.vision_model_output.attentions self.assertEqual(len(UpperCamelCase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _snake_case : str = to_atuple(vision_model.config.image_size ) _snake_case : Any = to_atuple(vision_model.config.patch_size ) _snake_case : List[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _snake_case : List[str] = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _snake_case : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(UpperCamelCase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def UpperCamelCase_ ( self : Optional[int] , UpperCamelCase : Optional[Any] , UpperCamelCase : str , UpperCamelCase : Optional[int] ): '''simple docstring''' pt_model.to(UpperCamelCase ) pt_model.eval() # prepare inputs _snake_case : List[Any] = inputs_dict _snake_case : Tuple = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): _snake_case : Tuple = pt_model(**UpperCamelCase ).to_tuple() _snake_case : Any = fx_model(**UpperCamelCase ).to_tuple() self.assertEqual(len(UpperCamelCase ) , len(UpperCamelCase ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(UpperCamelCase , pt_output.numpy() , 4e-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(UpperCamelCase ) _snake_case : Optional[int] = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase , from_pt=UpperCamelCase ) _snake_case : Any = fx_model_loaded(**UpperCamelCase ).to_tuple() self.assertEqual(len(UpperCamelCase ) , len(UpperCamelCase ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(UpperCamelCase , pt_output.numpy() , 4e-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(UpperCamelCase ) _snake_case : int = VisionTextDualEncoderModel.from_pretrained(UpperCamelCase , from_flax=UpperCamelCase ) pt_model_loaded.to(UpperCamelCase ) pt_model_loaded.eval() with torch.no_grad(): _snake_case : Tuple = pt_model_loaded(**UpperCamelCase ).to_tuple() self.assertEqual(len(UpperCamelCase ) , len(UpperCamelCase ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(UpperCamelCase , pt_output_loaded.numpy() , 4e-2 ) def UpperCamelCase_ ( self : Optional[Any] , UpperCamelCase : Tuple , UpperCamelCase : Dict , UpperCamelCase : Union[str, Any] ): '''simple docstring''' _snake_case : int = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase , UpperCamelCase ) _snake_case : Dict = VisionTextDualEncoderModel(UpperCamelCase ) _snake_case : List[Any] = FlaxVisionTextDualEncoderModel(UpperCamelCase ) _snake_case : Union[str, Any] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , UpperCamelCase ) _snake_case : List[Any] = fx_state self.check_pt_flax_equivalence(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def UpperCamelCase_ ( self : Optional[Any] , UpperCamelCase : int , UpperCamelCase : Dict , UpperCamelCase : List[str] ): '''simple docstring''' _snake_case : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase , UpperCamelCase ) _snake_case : Any = VisionTextDualEncoderModel(UpperCamelCase ) _snake_case : str = FlaxVisionTextDualEncoderModel(UpperCamelCase ) _snake_case : List[Any] = load_flax_weights_in_pytorch_model(UpperCamelCase , fx_model.params ) self.check_pt_flax_equivalence(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _snake_case : Optional[Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**UpperCamelCase ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _snake_case : List[str] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**UpperCamelCase ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _snake_case : Any = self.prepare_config_and_inputs() self.check_save_load(**UpperCamelCase ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _snake_case : Union[str, Any] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**UpperCamelCase ) @is_pt_flax_cross_test def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : List[str] = self.prepare_config_and_inputs() _snake_case : Any = config_inputs_dict.pop('vision_config' ) _snake_case : Union[str, Any] = config_inputs_dict.pop('text_config' ) _snake_case : Optional[Any] = config_inputs_dict self.check_equivalence_pt_to_flax(UpperCamelCase , UpperCamelCase , UpperCamelCase ) self.check_equivalence_flax_to_pt(UpperCamelCase , UpperCamelCase , UpperCamelCase ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case , _snake_case : int = self.get_pretrained_model_and_inputs() _snake_case : Dict = model_a(**UpperCamelCase ) _snake_case : Any = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(UpperCamelCase ) _snake_case : Any = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase ) _snake_case : int = model_a(**UpperCamelCase ) _snake_case : Any = after_outputs[0] _snake_case : List[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase , 1e-5 ) @require_flax class _lowerCAmelCase ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _snake_case : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-vit' , 'hf-internal-testing/tiny-bert' , vision_from_pt=UpperCamelCase , text_from_pt=UpperCamelCase , ) _snake_case : List[Any] = 13 _snake_case : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _snake_case : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _snake_case : List[Any] = random_attention_mask([batch_size, 4] ) _snake_case : List[str] = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def UpperCamelCase_ ( self : Optional[Any] , UpperCamelCase : str , UpperCamelCase : str ): '''simple docstring''' _snake_case : Optional[Any] = FlaxViTModel(UpperCamelCase ) _snake_case : Union[str, Any] = FlaxBertModel(UpperCamelCase ) return vision_model, text_model def UpperCamelCase_ ( self : int ): '''simple docstring''' _snake_case : Optional[int] = FlaxViTModelTester(self ) _snake_case : Optional[Any] = FlaxBertModelTester(self ) _snake_case : int = vit_model_tester.prepare_config_and_inputs() _snake_case : List[str] = bert_model_tester.prepare_config_and_inputs() _snake_case , _snake_case : Any = vision_config_and_inputs _snake_case , _snake_case , _snake_case , _snake_case : Any = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class _lowerCAmelCase ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : int ): '''simple docstring''' _snake_case : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-clip' , 'hf-internal-testing/tiny-bert' , vision_from_pt=UpperCamelCase , text_from_pt=UpperCamelCase , ) _snake_case : str = 13 _snake_case : Any = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _snake_case : Optional[int] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _snake_case : Union[str, Any] = random_attention_mask([batch_size, 4] ) _snake_case : Tuple = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def UpperCamelCase_ ( self : Optional[Any] , UpperCamelCase : Any , UpperCamelCase : Tuple ): '''simple docstring''' _snake_case : Tuple = FlaxCLIPVisionModel(UpperCamelCase ) _snake_case : Optional[int] = FlaxBertModel(UpperCamelCase ) return vision_model, text_model def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _snake_case : Any = FlaxCLIPVisionModelTester(self ) _snake_case : Tuple = FlaxBertModelTester(self ) _snake_case : Dict = clip_model_tester.prepare_config_and_inputs() _snake_case : Optional[Any] = bert_model_tester.prepare_config_and_inputs() _snake_case , _snake_case : Dict = vision_config_and_inputs _snake_case , _snake_case , _snake_case , _snake_case : Tuple = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _snake_case : List[str] = FlaxVisionTextDualEncoderModel.from_pretrained('clip-italian/clip-italian' , logit_scale_init_value=1.0 ) _snake_case : List[Any] = VisionTextDualEncoderProcessor.from_pretrained('clip-italian/clip-italian' ) _snake_case : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) _snake_case : Optional[Any] = processor( text=['una foto di un gatto', 'una foto di un cane'] , images=UpperCamelCase , padding=UpperCamelCase , return_tensors='np' ) _snake_case : Optional[Any] = model(**UpperCamelCase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _snake_case : Optional[int] = np.array([[1.2_28_47_27, 0.3_10_41_22]] ) self.assertTrue(np.allclose(outputs.logits_per_image , UpperCamelCase , atol=1e-3 ) )
260
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow lowerCAmelCase_ = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Tuple , UpperCamelCase : Path , UpperCamelCase : Union[str, None] = None , UpperCamelCase : Union[List[str], None] = None , UpperCamelCase : Union[str, List[str], None] = None , UpperCamelCase : bool = True , ): '''simple docstring''' _snake_case : List[str] = [file for file in os.listdir(UpperCamelCase ) if os.path.isfile(os.path.join(UpperCamelCase , UpperCamelCase ) )] if identifier is not None: _snake_case : Tuple = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(UpperCamelCase , UpperCamelCase ): for n_ in n_identifier: _snake_case : Dict = [file for file in files if n_ not in file] else: _snake_case : Optional[int] = [file for file in files if n_identifier not in file] _snake_case : List[Any] = ignore_files or [] ignore_files.append('__init__.py' ) _snake_case : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' , UpperCamelCase ) if only_modules: _snake_case : Any = file.split('.' )[0] try: _snake_case : List[Any] = getattr(UpperCamelCase , UpperCamelCase ) _snake_case : Any = doctest.DocTestSuite(UpperCamelCase ) _snake_case : Tuple = unittest.TextTestRunner().run(UpperCamelCase ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f"""{module_identifier} is not a module.""" ) else: _snake_case : Union[str, Any] = doctest.testfile(str('..' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _snake_case : Any = Path('src/transformers' ) _snake_case : Optional[int] = 'modeling' _snake_case : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(UpperCamelCase , identifier=UpperCamelCase , ignore_files=UpperCamelCase ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : str = Path('src/transformers' ) _snake_case : Dict = 'tokenization' self.analyze_directory(UpperCamelCase , identifier=UpperCamelCase ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _snake_case : str = Path('src/transformers' ) _snake_case : Optional[int] = 'configuration' self.analyze_directory(UpperCamelCase , identifier=UpperCamelCase ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _snake_case : Any = Path('src/transformers' ) _snake_case : List[str] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(UpperCamelCase , n_identifier=UpperCamelCase ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = Path('docs/source' ) _snake_case : Optional[Any] = ['favicon.ico'] self.analyze_directory(UpperCamelCase , ignore_files=UpperCamelCase , only_modules=UpperCamelCase )
260
1
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments SCREAMING_SNAKE_CASE_ : Any = logging.getLogger(__name__) @dataclass class a ( _lowerCAmelCase ): """simple docstring""" UpperCAmelCase = field( default=0.0, metadata={"help": "The label smoothing epsilon to apply (if not zero)."} ) UpperCAmelCase = field(default=_lowerCAmelCase, metadata={"help": "Whether to SortishSamler or not."} ) UpperCAmelCase = field( default=_lowerCAmelCase, metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) UpperCAmelCase = field(default=_lowerCAmelCase, metadata={"help": "whether to use adafactor"} ) UpperCAmelCase = field( default=_lowerCAmelCase, metadata={"help": "Encoder layer dropout probability. Goes into model.config."} ) UpperCAmelCase = field( default=_lowerCAmelCase, metadata={"help": "Decoder layer dropout probability. Goes into model.config."} ) UpperCAmelCase = field(default=_lowerCAmelCase, metadata={"help": "Dropout probability. Goes into model.config."} ) UpperCAmelCase = field( default=_lowerCAmelCase, metadata={"help": "Attention dropout probability. Goes into model.config."} ) UpperCAmelCase = field( default="linear", metadata={"help": F'Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'}, )
335
"""simple docstring""" from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo lowercase_ = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey\n and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin\n Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto\n Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and\n Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes\n and Jeffrey Dean},\n year={2016},\n eprint={1609.08144},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n' lowercase_ = '\\nThe BLEU score has some undesirable properties when used for single\nsentences, as it was designed to be a corpus measure. We therefore\nuse a slightly different score for our RL experiments which we call\nthe \'GLEU score\'. For the GLEU score, we record all sub-sequences of\n1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then\ncompute a recall, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the target (ground truth) sequence,\nand a precision, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the generated output sequence. Then\nGLEU score is simply the minimum of recall and precision. This GLEU\nscore\'s range is always between 0 (no matches) and 1 (all match) and\nit is symmetrical when switching output and target. According to\nour experiments, GLEU score correlates quite well with the BLEU\nmetric on a corpus level but does not have its drawbacks for our per\nsentence reward objective.\n' lowercase_ = '\\nComputes corpus-level Google BLEU (GLEU) score of translated segments against one or more references.\nInstead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching\ntokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values.\n\nArgs:\n predictions (list of str): list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references (list of list of str): list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n min_len (int): The minimum order of n-gram this function should extract. Defaults to 1.\n max_len (int): The maximum order of n-gram this function should extract. Defaults to 4.\n\nReturns:\n \'google_bleu\': google_bleu score\n\nExamples:\n Example 1:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.44\n\n Example 2:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.61\n\n Example 3:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2)\n >>> print(round(results["google_bleu"], 2))\n 0.53\n\n Example 4:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6)\n >>> print(round(results["google_bleu"], 2))\n 0.4\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''', id='''token''' ), id='''sequence''' ), '''references''': datasets.Sequence( datasets.Sequence(datasets.Value('''string''', id='''token''' ), id='''sequence''' ), id='''references''' ), } ), ) def _SCREAMING_SNAKE_CASE ( self : str, _lowerCamelCase : List[List[List[str]]], _lowerCamelCase : List[List[str]], _lowerCamelCase : int = 1, _lowerCamelCase : int = 4, ): '''simple docstring''' return { "google_bleu": gleu_score.corpus_gleu( list_of_references=_lowerCamelCase, hypotheses=_lowerCamelCase, min_len=_lowerCamelCase, max_len=_lowerCamelCase ) }
266
0
'''simple docstring''' def lowerCamelCase__ ( A : Tuple , A : str ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = len(__A ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None UpperCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__A ): return None UpperCAmelCase = sorted_collection[point] if current_item == item: return point else: if point < left: UpperCAmelCase = left UpperCAmelCase = point elif point > right: UpperCAmelCase = right UpperCAmelCase = point else: if item < current_item: UpperCAmelCase = point - 1 else: UpperCAmelCase = point + 1 return None def lowerCamelCase__ ( A : str , A : str , A : List[Any] , A : str ): '''simple docstring''' if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None UpperCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__A ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__A , __A , __A , __A ) elif point > right: return interpolation_search_by_recursion(__A , __A , __A , __A ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __A , __A , __A , point - 1 ) else: return interpolation_search_by_recursion( __A , __A , point + 1 , __A ) def lowerCamelCase__ ( A : Optional[Any] ): '''simple docstring''' if collection != sorted(__A ): raise ValueError('''Collection must be ascending sorted''' ) return True if __name__ == "__main__": import sys _lowercase : Any = 0 if debug == 1: _lowercase : Dict = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("""Sequence must be ascending sorted to apply interpolation search""") _lowercase : Dict = 67 _lowercase : Dict = interpolation_search(collection, target) if result is not None: print(F"""{target} found at positions: {result}""") else: print("""Not found""")
357
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase : int = logging.get_logger(__name__) _lowercase : Optional[Any] = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class UpperCamelCase__( lowerCAmelCase ): __magic_name__ : Union[str, Any] = "openai-gpt" __magic_name__ : Optional[int] = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , lowerCAmelCase : Optional[Any]=40478 , lowerCAmelCase : str=512 , lowerCAmelCase : List[Any]=768 , lowerCAmelCase : Tuple=12 , lowerCAmelCase : int=12 , lowerCAmelCase : List[str]="gelu" , lowerCAmelCase : List[str]=0.1 , lowerCAmelCase : List[Any]=0.1 , lowerCAmelCase : Dict=0.1 , lowerCAmelCase : Tuple=1E-5 , lowerCAmelCase : Tuple=0.02 , lowerCAmelCase : Optional[int]="cls_index" , lowerCAmelCase : Dict=True , lowerCAmelCase : Optional[int]=None , lowerCAmelCase : Dict=True , lowerCAmelCase : int=0.1 , **lowerCAmelCase : Optional[int] , )-> str: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = afn UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = attn_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = summary_type UpperCAmelCase = summary_use_proj UpperCAmelCase = summary_activation UpperCAmelCase = summary_first_dropout UpperCAmelCase = summary_proj_to_labels super().__init__(**lowerCAmelCase )
91
0
'''simple docstring''' import sys from collections import defaultdict class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Any ) -> str: UpperCAmelCase : Union[str, Any] = [] def A ( self : str , __snake_case : Any ) -> Optional[int]: return self.node_position[vertex] def A ( self : Dict , __snake_case : Optional[int] , __snake_case : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = pos def A ( self : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Dict , __snake_case : Dict , __snake_case : str ) -> Dict: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: UpperCAmelCase : Union[str, Any] = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: UpperCAmelCase : Tuple = 2 * start + 1 else: UpperCAmelCase : List[str] = 2 * start + 2 if heap[smallest_child] < heap[start]: UpperCAmelCase , UpperCAmelCase : str = heap[smallest_child], positions[smallest_child] UpperCAmelCase , UpperCAmelCase : str = ( heap[start], positions[start], ) UpperCAmelCase , UpperCAmelCase : Any = temp, tempa UpperCAmelCase : Union[str, Any] = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , __snake_case ) self.top_to_bottom(__snake_case , __snake_case , __snake_case , __snake_case ) def A ( self : Tuple , __snake_case : Any , __snake_case : Optional[Any] , __snake_case : Tuple , __snake_case : str ) -> Dict: UpperCAmelCase : Union[str, Any] = position[index] while index != 0: UpperCAmelCase : List[str] = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: UpperCAmelCase : List[Any] = heap[parent] UpperCAmelCase : str = position[parent] self.set_position(position[parent] , __snake_case ) else: UpperCAmelCase : Optional[Any] = val UpperCAmelCase : Dict = temp self.set_position(__snake_case , __snake_case ) break UpperCAmelCase : Optional[int] = parent else: UpperCAmelCase : Any = val UpperCAmelCase : Optional[int] = temp self.set_position(__snake_case , 0 ) def A ( self : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> Any: UpperCAmelCase : Tuple = len(__snake_case ) // 2 - 1 for i in range(__snake_case , -1 , -1 ): self.top_to_bottom(__snake_case , __snake_case , len(__snake_case ) , __snake_case ) def A ( self : Dict , __snake_case : Optional[Any] , __snake_case : int ) -> int: UpperCAmelCase : List[Any] = positions[0] UpperCAmelCase : Optional[int] = sys.maxsize self.top_to_bottom(__snake_case , 0 , len(__snake_case ) , __snake_case ) return temp def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: UpperCAmelCase : Union[str, Any] = Heap() UpperCAmelCase : str = [0] * len(_lowerCAmelCase ) UpperCAmelCase : str = [-1] * len(_lowerCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph UpperCAmelCase : Optional[int] = [] # Heap of Distance of vertices from their neighboring vertex UpperCAmelCase : List[str] = [] for vertex in range(len(_lowerCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(_lowerCAmelCase ) heap.node_position.append(_lowerCAmelCase ) UpperCAmelCase : List[str] = [] UpperCAmelCase : Tuple = 1 UpperCAmelCase : Union[str, Any] = sys.maxsize for neighbor, distance in adjacency_list[0]: UpperCAmelCase : List[Any] = 0 UpperCAmelCase : Tuple = distance heap.heapify(_lowerCAmelCase , _lowerCAmelCase ) for _ in range(1 , len(_lowerCAmelCase ) ): UpperCAmelCase : str = heap.delete_minimum(_lowerCAmelCase , _lowerCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) UpperCAmelCase : Optional[Any] = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(_lowerCAmelCase )] ): UpperCAmelCase : Tuple = distance heap.bottom_to_top( _lowerCAmelCase , heap.get_position(_lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > UpperCamelCase__: Any = int(input("Enter number of edges: ").strip()) UpperCamelCase__: Any = defaultdict(list) for _ in range(edges_number): UpperCamelCase__: Tuple = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
23
"""simple docstring""" import mpmath # for roots of unity import numpy as np class snake_case_: def __init__( self : str , UpperCamelCase_ : int=None , UpperCamelCase_ : List[str]=None ): # Input as list lowerCAmelCase : str = list(poly_a or [0] )[:] lowerCAmelCase : Any = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() lowerCAmelCase : Optional[int] = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() lowerCAmelCase : Union[str, Any] = len(self.polyB ) # Add 0 to make lengths equal a power of 2 lowerCAmelCase : str = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform lowerCAmelCase : int = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product lowerCAmelCase : int = self.__multiply() def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : str ): lowerCAmelCase : Optional[Any] = [[x] for x in self.polyA] if which == '''A''' else [[x] for x in self.polyB] # Corner case if len(UpperCamelCase_ ) <= 1: return dft[0] # lowerCAmelCase : Tuple = self.c_max_length // 2 while next_ncol > 0: lowerCAmelCase : Dict = [[] for i in range(UpperCamelCase_ )] lowerCAmelCase : List[Any] = self.root**next_ncol # First half of next step lowerCAmelCase : Dict = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(UpperCamelCase_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step lowerCAmelCase : int = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(UpperCamelCase_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update lowerCAmelCase : Optional[Any] = new_dft lowerCAmelCase : Union[str, Any] = next_ncol // 2 return dft[0] def lowerCamelCase__ ( self : List[Any] ): lowerCAmelCase : Optional[Any] = self.__dft('''A''' ) lowerCAmelCase : Optional[int] = self.__dft('''B''' ) lowerCAmelCase : Any = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT lowerCAmelCase : str = 2 while next_ncol <= self.c_max_length: lowerCAmelCase : Union[str, Any] = [[] for i in range(UpperCamelCase_ )] lowerCAmelCase : Optional[Any] = self.root ** (next_ncol // 2) lowerCAmelCase : Tuple = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update lowerCAmelCase : Any = new_inverse_c next_ncol *= 2 # Unpack lowerCAmelCase : Optional[int] = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : int ): lowerCAmelCase : int = '''A = ''' + ''' + '''.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) lowerCAmelCase : str = '''B = ''' + ''' + '''.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) lowerCAmelCase : int = '''A*B = ''' + ''' + '''.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return F'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
60
0
"""simple docstring""" lowercase__ = 65521 def __lowerCamelCase ( __UpperCamelCase ) -> int: """simple docstring""" lowerCAmelCase_ : Optional[int] = 1 lowerCAmelCase_ : Optional[Any] = 0 for plain_chr in plain_text: lowerCAmelCase_ : str = (a + ord(__UpperCamelCase )) % MOD_ADLER lowerCAmelCase_ : str = (b + a) % MOD_ADLER return (b << 16) | a
161
"""simple docstring""" import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """vocab.txt"""} lowercase__ = { """vocab_file""": { """openbmb/cpm-ant-10b""": """https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt""", }, } lowercase__ = { """openbmb/cpm-ant-10b""": 1024, } def __lowerCamelCase ( __UpperCamelCase ) -> Dict: """simple docstring""" lowerCAmelCase_ : int = collections.OrderedDict() with open(__UpperCamelCase , "r" , encoding="utf-8" ) as reader: lowerCAmelCase_ : List[Any] = reader.readlines() for index, token in enumerate(__UpperCamelCase ): lowerCAmelCase_ : List[str] = token.rstrip("\n" ) lowerCAmelCase_ : str = index return vocab class __lowerCamelCase ( A__ ): '''simple docstring''' def __init__( self : Tuple , a_ : Dict , a_ : Optional[Any]="<unk>" , a_ : List[str]=2_00 ): lowerCAmelCase_ : int = vocab lowerCAmelCase_ : List[Any] = unk_token lowerCAmelCase_ : List[Any] = max_input_chars_per_word def lowerCamelCase ( self : Any , a_ : Optional[int] ): lowerCAmelCase_ : Optional[int] = list(a_ ) if len(a_ ) > self.max_input_chars_per_word: return [self.unk_token] lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = [] while start < len(a_ ): lowerCAmelCase_ : Any = len(a_ ) lowerCAmelCase_ : Any = None while start < end: lowerCAmelCase_ : Union[str, Any] = "".join(chars[start:end] ) if substr in self.vocab: lowerCAmelCase_ : Any = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(a_ ) lowerCAmelCase_ : int = end return sub_tokens class __lowerCamelCase ( A__ ): '''simple docstring''' a_ : str = VOCAB_FILES_NAMES a_ : int = PRETRAINED_VOCAB_FILES_MAP a_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : List[Any] = ["""input_ids""", """attention_mask"""] a_ : Union[str, Any] = False def __init__( self : Union[str, Any] , a_ : List[str] , a_ : Dict="<d>" , a_ : Tuple="</d>" , a_ : Tuple="<s>" , a_ : int="</s>" , a_ : Tuple="<pad>" , a_ : Dict="<unk>" , a_ : Any="</n>" , a_ : Optional[int]="</_>" , a_ : List[Any]="left" , **a_ : List[Any] , ): requires_backends(self , ["jieba"] ) super().__init__( bod_token=a_ , eod_token=a_ , bos_token=a_ , eos_token=a_ , pad_token=a_ , unk_token=a_ , line_token=a_ , space_token=a_ , padding_side=a_ , **a_ , ) lowerCAmelCase_ : Optional[Any] = bod_token lowerCAmelCase_ : Union[str, Any] = eod_token lowerCAmelCase_ : Optional[Any] = load_vocab(a_ ) lowerCAmelCase_ : List[str] = self.encoder[space_token] lowerCAmelCase_ : int = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] lowerCAmelCase_ : Tuple = collections.OrderedDict(sorted(self.encoder.items() , key=lambda a_ : x[1] ) ) lowerCAmelCase_ : Any = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : str = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def lowerCamelCase ( self : List[Any] ): return self.encoder[self.bod_token] @property def lowerCamelCase ( self : List[str] ): return self.encoder[self.eod_token] @property def lowerCamelCase ( self : int ): return self.encoder["\n"] @property def lowerCamelCase ( self : Tuple ): return len(self.encoder ) def lowerCamelCase ( self : Optional[int] ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase ( self : Optional[int] , a_ : Any ): lowerCAmelCase_ : Optional[int] = [] for x in jieba.cut(a_ , cut_all=a_ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(a_ ) ) return output_tokens def lowerCamelCase ( self : Optional[Any] , a_ : List[str] , **a_ : Tuple ): lowerCAmelCase_ : List[Any] = [i for i in token_ids if i >= 0] lowerCAmelCase_ : List[Any] = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(a_ , **a_ ) def lowerCamelCase ( self : Optional[Any] , a_ : Union[str, Any] ): return token in self.encoder def lowerCamelCase ( self : List[Any] , a_ : List[str] ): return "".join(a_ ) def lowerCamelCase ( self : Union[str, Any] , a_ : str ): return self.encoder.get(a_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase ( self : Union[str, Any] , a_ : int ): return self.decoder.get(a_ , self.unk_token ) def lowerCamelCase ( self : List[str] , a_ : str , a_ : Optional[str] = None ): if os.path.isdir(a_ ): lowerCAmelCase_ : Union[str, Any] = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowerCAmelCase_ : Tuple = (filename_prefix + "-" if filename_prefix else "") + save_directory lowerCAmelCase_ : str = 0 if " " in self.encoder: lowerCAmelCase_ : Optional[int] = self.encoder[" "] del self.encoder[" "] if "\n" in self.encoder: lowerCAmelCase_ : Optional[int] = self.encoder["\n"] del self.encoder["\n"] lowerCAmelCase_ : int = collections.OrderedDict(sorted(self.encoder.items() , key=lambda a_ : x[1] ) ) with open(a_ , "w" , encoding="utf-8" ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(token + "\n" ) index += 1 return (vocab_file,) def lowerCamelCase ( self : int , a_ : List[int] , a_ : List[int] = None ): if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def lowerCamelCase ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a_ , token_ids_a=a_ , already_has_special_tokens=a_ ) if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) return [1] + ([0] * len(a_ ))
161
1
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1_024 , lowerCAmelCase_=1_024 , lowerCAmelCase_=False , **lowerCAmelCase_ )-> Any: lowerCAmelCase_ : int = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) lowerCAmelCase_ : Dict = SeqaSeqDataset(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , type_path='''train''' , **lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = tok.pad_token_id def get_lens(lowerCAmelCase_ ): lowerCAmelCase_ : int = tqdm( DataLoader(lowerCAmelCase_ , batch_size=512 , num_workers=8 , shuffle=lowerCAmelCase_ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) lowerCAmelCase_ : str = [] for batch in dl: lowerCAmelCase_ : str = batch['''input_ids'''].ne(lowerCAmelCase_ ).sum(1 ).tolist() lowerCAmelCase_ : Optional[int] = batch['''labels'''].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_ , lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_ , lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens lowerCAmelCase_ : Tuple = get_lens(lowerCAmelCase_ ) lowerCAmelCase_ : int = SeqaSeqDataset(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , type_path='''val''' , **lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_ , train_ds.len_file ) pickle_save(lowerCAmelCase_ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
262
from __future__ import annotations import requests def lowerCAmelCase ( lowerCAmelCase_ )-> dict: lowerCAmelCase_ : List[Any] = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> list[dict]: lowerCAmelCase_ : List[Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCAmelCase_ : Tuple = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> str: lowerCAmelCase_ : Optional[Any] = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join('''* [{title}]({url})'''.format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
262
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _lowerCAmelCase ( A__: List[str]=None ): '''simple docstring''' if subparsers is not None: UpperCAmelCase = subparsers.add_parser('''test''' ) else: UpperCAmelCase = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=A__ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=A__ ) return parser def _lowerCAmelCase ( A__: List[str] ): '''simple docstring''' UpperCAmelCase = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: UpperCAmelCase = script_name else: UpperCAmelCase = F"""--config_file={args.config_file} {script_name}""" UpperCAmelCase = ['''accelerate-launch'''] + test_args.split() UpperCAmelCase = execute_subprocess_async(A__ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = test_command_parser() UpperCAmelCase = parser.parse_args() test_command(A__ ) if __name__ == "__main__": main()
152
import string def _lowerCAmelCase ( A__: str ): '''simple docstring''' for key in range(len(string.ascii_uppercase ) ): UpperCAmelCase = '''''' for symbol in message: if symbol in string.ascii_uppercase: UpperCAmelCase = string.ascii_uppercase.find(A__ ) UpperCAmelCase = num - key if num < 0: UpperCAmelCase = num + len(string.ascii_uppercase ) UpperCAmelCase = translated + string.ascii_uppercase[num] else: UpperCAmelCase = translated + symbol print(F"""Decryption using Key #{key}: {translated}""" ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = input('''Encrypted message: ''' ) UpperCAmelCase = message.upper() decrypt(A__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
152
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ : List[str] = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : List[Any] = [ 'XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaXLForCausalLM', 'XLMRobertaXLForMaskedLM', 'XLMRobertaXLForMultipleChoice', 'XLMRobertaXLForQuestionAnswering', 'XLMRobertaXLForSequenceClassification', 'XLMRobertaXLForTokenClassification', 'XLMRobertaXLModel', 'XLMRobertaXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys UpperCAmelCase_ : int = _LazyModule(__name__, globals()['__file__'], _import_structure)
32
def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : Union[str, Any] = abs(snake_case__ ) __UpperCAmelCase : Dict = 0 while n > 0: res += n % 10 n //= 10 return res def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : Tuple = abs(snake_case__ ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def _UpperCamelCase ( snake_case__ ) -> int: return sum(int(snake_case__ ) for c in str(abs(snake_case__ ) ) ) def _UpperCamelCase ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(snake_case__, snake_case__ ) -> None: __UpperCAmelCase : Union[str, Any] = f'''{func.__name__}({value})''' __UpperCAmelCase : List[str] = timeit(f'''__main__.{call}''', setup="import __main__" ) print(f'''{call:56} = {func(snake_case__ )} -- {timing:.4f} seconds''' ) for value in (26_2144, 1125_8999_0684_2624, 126_7650_6002_2822_9401_4967_0320_5376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(snake_case__, snake_case__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
157
0
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __UpperCamelCase : Union[str, Any] = 'src/diffusers' __UpperCamelCase : Optional[int] = '.' # This is to make sure the diffusers module imported is the one in the repo. __UpperCamelCase : Union[str, Any] = importlib.util.spec_from_file_location( 'diffusers', os.path.join(DIFFUSERS_PATH, '__init__.py'), submodule_search_locations=[DIFFUSERS_PATH], ) __UpperCamelCase : Any = spec.loader.load_module() def A ( _lowercase , _lowercase ): return line.startswith(_lowercase ) or len(_lowercase ) <= 1 or re.search(R'''^\s*\)(\s*->.*:|:)\s*$''' , _lowercase ) is not None def A ( _lowercase ): SCREAMING_SNAKE_CASE : Tuple = object_name.split('''.''' ) SCREAMING_SNAKE_CASE : int = 0 # First let's find the module where our object lives. SCREAMING_SNAKE_CASE : List[Any] = parts[i] while i < len(_lowercase ) and not os.path.isfile(os.path.join(_lowercase , f"""{module}.py""" ) ): i += 1 if i < len(_lowercase ): SCREAMING_SNAKE_CASE : Dict = os.path.join(_lowercase , parts[i] ) if i >= len(_lowercase ): raise ValueError(f"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(_lowercase , f"""{module}.py""" ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: SCREAMING_SNAKE_CASE : Optional[int] = f.readlines() # Now let's find the class / func in the code! SCREAMING_SNAKE_CASE : Dict = '''''' SCREAMING_SNAKE_CASE : Dict = 0 for name in parts[i + 1 :]: while ( line_index < len(_lowercase ) and re.search(Rf"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(_lowercase ): raise ValueError(f""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). SCREAMING_SNAKE_CASE : str = line_index while line_index < len(_lowercase ) and _should_continue(lines[line_index] , _lowercase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 SCREAMING_SNAKE_CASE : Union[str, Any] = lines[start_index:line_index] return "".join(_lowercase ) __UpperCamelCase : Optional[int] = re.compile(R'^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)') __UpperCamelCase : Any = re.compile(R'^\s*(\S+)->(\S+)(\s+.*|$)') __UpperCamelCase : str = re.compile(R'<FILL\s+[^>]*>') def A ( _lowercase ): SCREAMING_SNAKE_CASE : List[Any] = code.split('''\n''' ) SCREAMING_SNAKE_CASE : List[str] = 0 while idx < len(_lowercase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(_lowercase ): return re.search(R'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def A ( _lowercase ): SCREAMING_SNAKE_CASE : Dict = len(get_indent(_lowercase ) ) > 0 if has_indent: SCREAMING_SNAKE_CASE : List[str] = f"""class Bla:\n{code}""" SCREAMING_SNAKE_CASE : Dict = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=_lowercase ) SCREAMING_SNAKE_CASE : List[Any] = black.format_str(_lowercase , mode=_lowercase ) SCREAMING_SNAKE_CASE : Tuple = style_docstrings_in_code(_lowercase ) return result[len('''class Bla:\n''' ) :] if has_indent else result def A ( _lowercase , _lowercase=False ): with open(_lowercase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: SCREAMING_SNAKE_CASE : Tuple = f.readlines() SCREAMING_SNAKE_CASE : List[Any] = [] SCREAMING_SNAKE_CASE : Tuple = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(_lowercase ): SCREAMING_SNAKE_CASE : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. SCREAMING_SNAKE_CASE : Optional[int] = search.groups() SCREAMING_SNAKE_CASE : Optional[Any] = find_code_in_diffusers(_lowercase ) SCREAMING_SNAKE_CASE : List[Any] = get_indent(_lowercase ) SCREAMING_SNAKE_CASE : int = line_index + 1 if indent == theoretical_indent else line_index + 2 SCREAMING_SNAKE_CASE : str = theoretical_indent SCREAMING_SNAKE_CASE : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. SCREAMING_SNAKE_CASE : List[str] = True while line_index < len(_lowercase ) and should_continue: line_index += 1 if line_index >= len(_lowercase ): break SCREAMING_SNAKE_CASE : Dict = lines[line_index] SCREAMING_SNAKE_CASE : Any = _should_continue(_lowercase , _lowercase ) and re.search(f"""^{indent}# End copy""" , _lowercase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 SCREAMING_SNAKE_CASE : Any = lines[start_index:line_index] SCREAMING_SNAKE_CASE : Tuple = ''''''.join(_lowercase ) # Remove any nested `Copied from` comments to avoid circular copies SCREAMING_SNAKE_CASE : Optional[Any] = [line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(_lowercase ) is None] SCREAMING_SNAKE_CASE : List[str] = '''\n'''.join(_lowercase ) # Before comparing, use the `replace_pattern` on the original code. if len(_lowercase ) > 0: SCREAMING_SNAKE_CASE : Tuple = replace_pattern.replace('''with''' , '''''' ).split(''',''' ) SCREAMING_SNAKE_CASE : int = [_re_replace_pattern.search(_lowercase ) for p in patterns] for pattern in patterns: if pattern is None: continue SCREAMING_SNAKE_CASE : List[Any] = pattern.groups() SCREAMING_SNAKE_CASE : List[str] = re.sub(_lowercase , _lowercase , _lowercase ) if option.strip() == "all-casing": SCREAMING_SNAKE_CASE : Optional[int] = re.sub(obja.lower() , obja.lower() , _lowercase ) SCREAMING_SNAKE_CASE : List[str] = re.sub(obja.upper() , obja.upper() , _lowercase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line SCREAMING_SNAKE_CASE : Tuple = blackify(lines[start_index - 1] + theoretical_code ) SCREAMING_SNAKE_CASE : Optional[int] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: SCREAMING_SNAKE_CASE : Tuple = lines[:start_index] + [theoretical_code] + lines[line_index:] SCREAMING_SNAKE_CASE : int = start_index + 1 if overwrite and len(_lowercase ) > 0: # Warn the user a file has been modified. print(f"""Detected changes, rewriting {filename}.""" ) with open(_lowercase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowercase ) return diffs def A ( _lowercase = False ): SCREAMING_SNAKE_CASE : Dict = glob.glob(os.path.join(_lowercase , '''**/*.py''' ) , recursive=_lowercase ) SCREAMING_SNAKE_CASE : List[str] = [] for filename in all_files: SCREAMING_SNAKE_CASE : Union[str, Any] = is_copy_consistent(_lowercase , _lowercase ) diffs += [f"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(_lowercase ) > 0: SCREAMING_SNAKE_CASE : Optional[int] = '''\n'''.join(_lowercase ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": __UpperCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') __UpperCamelCase : List[Any] = parser.parse_args() check_copies(args.fix_and_overwrite)
354
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class lowercase__ : def __init__( self : Any , UpperCamelCase__ : Any , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : List[Any]="resnet50" , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[Any]=32 , UpperCamelCase__ : List[Any]=3 , UpperCamelCase__ : Any=True , UpperCamelCase__ : int=True , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = parent SCREAMING_SNAKE_CASE : Union[str, Any] = out_indices if out_indices is not None else [4] SCREAMING_SNAKE_CASE : List[Any] = stage_names SCREAMING_SNAKE_CASE : int = out_features SCREAMING_SNAKE_CASE : Optional[int] = backbone SCREAMING_SNAKE_CASE : Union[str, Any] = batch_size SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : List[Any] = use_pretrained_backbone SCREAMING_SNAKE_CASE : Dict = is_training def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : Tuple = self.get_config() return config, pixel_values def __A ( self : List[Any] ): '''simple docstring''' return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def __A ( self : List[str] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = TimmBackbone(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : Dict = model(UpperCamelCase__ ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = config_and_inputs SCREAMING_SNAKE_CASE : Tuple = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch @require_timm class lowercase__ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase): UpperCamelCase_ = (TimmBackbone,) if is_torch_available() else () UpperCamelCase_ = {"""feature-extraction""": TimmBackbone} if is_torch_available() else {} UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = TimmBackboneModelTester(self ) SCREAMING_SNAKE_CASE : Tuple = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def __A ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '''resnet18''' SCREAMING_SNAKE_CASE : str = '''microsoft/resnet-18''' SCREAMING_SNAKE_CASE : Dict = AutoBackbone.from_pretrained(UpperCamelCase__ , use_timm_backbone=UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = AutoBackbone.from_pretrained(UpperCamelCase__ ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) SCREAMING_SNAKE_CASE : List[str] = AutoBackbone.from_pretrained(UpperCamelCase__ , use_timm_backbone=UpperCamelCase__ , out_indices=[1, 2, 3] ) SCREAMING_SNAKE_CASE : Optional[Any] = AutoBackbone.from_pretrained(UpperCamelCase__ , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip('''TimmBackbone doesn\'t support feed forward chunking''' ) def __A ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t have num_hidden_layers attribute''' ) def __A ( self : int ): '''simple docstring''' pass @unittest.skip('''TimmBackbone initialization is managed on the timm side''' ) def __A ( self : Optional[Any] ): '''simple docstring''' pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def __A ( self : List[Any] ): '''simple docstring''' pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def __A ( self : Any ): '''simple docstring''' pass @unittest.skip('''TimmBackbone model cannot be created without specifying a backbone checkpoint''' ) def __A ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def __A ( self : int ): '''simple docstring''' pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def __A ( self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def __A ( self : Optional[Any] ): '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def __A ( self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def __A ( self : List[str] ): '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t have hidden size info in its configuration.''' ) def __A ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t support output_attentions.''' ) def __A ( self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip('''Safetensors is not supported by timm.''' ) def __A ( self : List[Any] ): '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __A ( self : int ): '''simple docstring''' pass def __A ( self : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Dict = model_class(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE : str = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE : Optional[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE : int = True SCREAMING_SNAKE_CASE : Any = self.has_attentions # no need to test all models as different heads yield the same functionality SCREAMING_SNAKE_CASE : Any = self.all_model_classes[0] SCREAMING_SNAKE_CASE : List[str] = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Any = model(**UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = outputs[0][-1] # Encoder-/Decoder-only models SCREAMING_SNAKE_CASE : List[Any] = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: SCREAMING_SNAKE_CASE : Any = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=UpperCamelCase__ ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Tuple = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() SCREAMING_SNAKE_CASE : Optional[Any] = model(**UpperCamelCase__ ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None SCREAMING_SNAKE_CASE : List[str] = copy.deepcopy(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = None SCREAMING_SNAKE_CASE : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() SCREAMING_SNAKE_CASE : List[str] = model(**UpperCamelCase__ ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights SCREAMING_SNAKE_CASE : Optional[Any] = copy.deepcopy(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() SCREAMING_SNAKE_CASE : int = model(**UpperCamelCase__ )
258
0
'''simple docstring''' import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } __a = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } __a = {"facebook/blenderbot_small-90M": 512} def __snake_case( _lowerCAmelCase ) -> Optional[Any]: snake_case__ : Any = set() snake_case__ : Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case__ : str = char snake_case__ : Dict = set(_lowerCAmelCase ) return pairs class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ["input_ids", "attention_mask"] def __init__( self : str , snake_case_ : Optional[int] , snake_case_ : Optional[int] , snake_case_ : int="__start__" , snake_case_ : Optional[int]="__end__" , snake_case_ : str="__unk__" , snake_case_ : List[Any]="__null__" , **snake_case_ : int , ): super().__init__(unk_token=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , pad_token=snake_case_ , **snake_case_ ) with open(snake_case_ , encoding="""utf-8""" ) as vocab_handle: snake_case__ : Dict = json.load(snake_case_ ) snake_case__ : Optional[Any] = {v: k for k, v in self.encoder.items()} with open(snake_case_ , encoding="""utf-8""" ) as merges_handle: snake_case__ : str = merges_handle.read().split("""\n""" )[1:-1] snake_case__ : Tuple = [tuple(merge.split() ) for merge in merges] snake_case__ : Any = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) snake_case__ : Dict = {} @property def lowerCamelCase ( self : str ): return len(self.encoder ) def lowerCamelCase ( self : List[Any] ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase ( self : Dict , snake_case_ : str ): if token in self.cache: return self.cache[token] snake_case__ : Tuple = re.sub("""([.,!?()])""" , r""" \1""" , snake_case_ ) snake_case__ : Optional[int] = re.sub("""(')""" , r""" \1 """ , snake_case_ ) snake_case__ : int = re.sub(r"""\s{2,}""" , """ """ , snake_case_ ) if "\n" in token: snake_case__ : int = token.replace("""\n""" , """ __newln__""" ) snake_case__ : Dict = token.split(""" """ ) snake_case__ : Union[str, Any] = [] for token in tokens: if not len(snake_case_ ): continue snake_case__ : List[str] = token.lower() snake_case__ : Dict = tuple(snake_case_ ) snake_case__ : Optional[Any] = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) snake_case__ : int = get_pairs(snake_case_ ) if not pairs: words.append(snake_case_ ) continue while True: snake_case__ : Tuple = min(snake_case_ , key=lambda snake_case_ : self.bpe_ranks.get(snake_case_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break snake_case__ , snake_case__ : List[Any] = bigram snake_case__ : str = [] snake_case__ : Tuple = 0 while i < len(snake_case_ ): try: snake_case__ : Union[str, Any] = word.index(snake_case_ , snake_case_ ) new_word.extend(word[i:j] ) snake_case__ : List[Any] = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(snake_case_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case__ : int = tuple(snake_case_ ) snake_case__ : Optional[Any] = new_word if len(snake_case_ ) == 1: break else: snake_case__ : str = get_pairs(snake_case_ ) snake_case__ : Optional[int] = """@@ """.join(snake_case_ ) snake_case__ : Any = word[:-4] snake_case__ : str = word words.append(snake_case_ ) return " ".join(snake_case_ ) def lowerCamelCase ( self : Tuple , snake_case_ : str ): snake_case__ : Tuple = [] snake_case__ : str = re.findall(r"""\S+\n?""" , snake_case_ ) for token in words: split_tokens.extend(list(self.bpe(snake_case_ ).split(""" """ ) ) ) return split_tokens def lowerCamelCase ( self : Optional[Any] , snake_case_ : str ): snake_case__ : Dict = token.lower() return self.encoder.get(snake_case_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase ( self : int , snake_case_ : int ): return self.decoder.get(snake_case_ , self.unk_token ) def lowerCamelCase ( self : Any , snake_case_ : List[str] ): snake_case__ : List[str] = """ """.join(snake_case_ ).replace("""@@ """ , """""" ).strip() return out_string def lowerCamelCase ( self : Optional[Any] , snake_case_ : str , snake_case_ : Optional[str] = None ): if not os.path.isdir(snake_case_ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return snake_case__ : Union[str, Any] = os.path.join( snake_case_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) snake_case__ : List[Any] = os.path.join( snake_case_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(snake_case_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case_ , ensure_ascii=snake_case_ ) + """\n""" ) snake_case__ : Optional[Any] = 0 with open(snake_case_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case_ : kv[1] ): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." """ Please check that the tokenizer is not corrupted!""" ) snake_case__ : int = token_index writer.write(""" """.join(snake_case_ ) + """\n""" ) index += 1 return vocab_file, merge_file
35
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : int = { 'Helsinki-NLP/opus-mt-en-de': 'https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json', # See all Marian models at https://huggingface.co/models?filter=marian } class _UpperCAmelCase ( lowercase_ ): UpperCamelCase = '''marian''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self :int , __UpperCamelCase :Any=5_81_01 , __UpperCamelCase :int=None , __UpperCamelCase :Union[str, Any]=10_24 , __UpperCamelCase :Union[str, Any]=12 , __UpperCamelCase :str=40_96 , __UpperCamelCase :int=16 , __UpperCamelCase :int=12 , __UpperCamelCase :Optional[Any]=40_96 , __UpperCamelCase :Optional[Any]=16 , __UpperCamelCase :Dict=0.0 , __UpperCamelCase :Dict=0.0 , __UpperCamelCase :str=True , __UpperCamelCase :Optional[int]=True , __UpperCamelCase :Any="gelu" , __UpperCamelCase :Any=10_24 , __UpperCamelCase :List[Any]=0.1 , __UpperCamelCase :Optional[Any]=0.0 , __UpperCamelCase :Union[str, Any]=0.0 , __UpperCamelCase :Tuple=0.02 , __UpperCamelCase :List[str]=5_81_00 , __UpperCamelCase :str=False , __UpperCamelCase :Optional[int]=5_81_00 , __UpperCamelCase :List[Any]=0 , __UpperCamelCase :List[str]=0 , __UpperCamelCase :Dict=True , **__UpperCamelCase :Tuple , ): A = vocab_size A = decoder_vocab_size or vocab_size A = max_position_embeddings A = d_model A = encoder_ffn_dim A = encoder_layers A = encoder_attention_heads A = decoder_ffn_dim A = decoder_layers A = decoder_attention_heads A = dropout A = attention_dropout A = activation_dropout A = activation_function A = init_std A = encoder_layerdrop A = decoder_layerdrop A = use_cache A = encoder_layers A = scale_embedding # scale factor will be sqrt(d_model) if True A = share_encoder_decoder_embeddings super().__init__( pad_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , is_encoder_decoder=__UpperCamelCase , decoder_start_token_id=__UpperCamelCase , forced_eos_token_id=__UpperCamelCase , **__UpperCamelCase , ) class _UpperCAmelCase ( lowercase_ ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def lowerCamelCase ( self :List[str] ): if self.task in ["default", "seq2seq-lm"]: A = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: A = {0: "batch"} A = {0: "batch", 1: "past_decoder_sequence + sequence"} else: A = {0: "batch", 1: "decoder_sequence"} A = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__UpperCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. A = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: A, A = self.num_layers for i in range(__UpperCamelCase ): A = {0: "batch", 2: "past_sequence + sequence"} A = {0: "batch", 2: "past_sequence + sequence"} else: A = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def lowerCamelCase ( self :List[str] ): if self.task in ["default", "seq2seq-lm"]: A = super().outputs else: A = super(__UpperCamelCase , self ).outputs if self.use_past: A, A = self.num_layers for i in range(__UpperCamelCase ): A = {0: "batch", 2: "past_sequence + sequence"} A = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def lowerCamelCase ( self :Optional[int] , __UpperCamelCase :PreTrainedTokenizer , __UpperCamelCase :int = -1 , __UpperCamelCase :int = -1 , __UpperCamelCase :bool = False , __UpperCamelCase :Optional[TensorType] = None , ): A = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Generate decoder inputs A = seq_length if not self.use_past else 1 A = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) A = {f"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} A = dict(**__UpperCamelCase , **__UpperCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A, A = common_inputs["input_ids"].shape A = common_inputs["decoder_input_ids"].shape[1] A, A = self.num_attention_heads A = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) A = decoder_seq_length + 3 A = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) A = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(__UpperCamelCase , __UpperCamelCase )] , dim=1 ) A = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered A, A = self.num_layers A = min(__UpperCamelCase , __UpperCamelCase ) A = max(__UpperCamelCase , __UpperCamelCase ) - min_num_layers A = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(__UpperCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), ) ) # TODO: test this. A = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(__UpperCamelCase , __UpperCamelCase ): common_inputs["past_key_values"].append((torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase )) ) return common_inputs def lowerCamelCase ( self :Optional[int] , __UpperCamelCase :PreTrainedTokenizer , __UpperCamelCase :int = -1 , __UpperCamelCase :int = -1 , __UpperCamelCase :bool = False , __UpperCamelCase :Optional[TensorType] = None , ): A = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A, A = common_inputs["input_ids"].shape # Not using the same length for past_key_values A = seqlen + 2 A, A = self.num_layers A, A = self.num_attention_heads A = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) A = common_inputs["attention_mask"].dtype A = torch.cat( [common_inputs["attention_mask"], torch.ones(__UpperCamelCase , __UpperCamelCase , dtype=__UpperCamelCase )] , dim=1 ) A = [ (torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase )) for _ in range(__UpperCamelCase ) ] return common_inputs def lowerCamelCase ( self :Tuple , __UpperCamelCase :PreTrainedTokenizer , __UpperCamelCase :int = -1 , __UpperCamelCase :int = -1 , __UpperCamelCase :bool = False , __UpperCamelCase :Optional[TensorType] = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX A = compute_effective_axis_dimension( __UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX A = tokenizer.num_special_tokens_to_add(__UpperCamelCase ) A = compute_effective_axis_dimension( __UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__UpperCamelCase ) # Generate dummy inputs according to compute batch and sequence A = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size A = dict(tokenizer(__UpperCamelCase , return_tensors=__UpperCamelCase ) ) return common_inputs def lowerCamelCase ( self :List[Any] , __UpperCamelCase :PreTrainedTokenizer , __UpperCamelCase :int = -1 , __UpperCamelCase :int = -1 , __UpperCamelCase :bool = False , __UpperCamelCase :Optional[TensorType] = None , ): if self.task in ["default", "seq2seq-lm"]: A = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __UpperCamelCase , batch_size=__UpperCamelCase , seq_length=__UpperCamelCase , is_pair=__UpperCamelCase , framework=__UpperCamelCase ) else: A = self._generate_dummy_inputs_for_causal_lm( __UpperCamelCase , batch_size=__UpperCamelCase , seq_length=__UpperCamelCase , is_pair=__UpperCamelCase , framework=__UpperCamelCase ) return common_inputs def lowerCamelCase ( self :List[Any] , __UpperCamelCase :Tuple , __UpperCamelCase :List[str] , __UpperCamelCase :str , __UpperCamelCase :str ): if self.task in ["default", "seq2seq-lm"]: A = super()._flatten_past_key_values_(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: A = super(__UpperCamelCase , self )._flatten_past_key_values_( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) @property def lowerCamelCase ( self :List[str] ): return 1e-4
292
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowercase : """simple docstring""" def __init__( self : Tuple , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any]=3 , lowerCAmelCase__ : Optional[Any]=32 , lowerCAmelCase__ : Union[str, Any]=3 , lowerCAmelCase__ : List[str]=10 , lowerCAmelCase__ : List[Any]=[10, 20, 30, 40] , lowerCAmelCase__ : int=[1, 1, 2, 1] , lowerCAmelCase__ : Any=True , lowerCAmelCase__ : Any=True , lowerCAmelCase__ : Tuple="relu" , lowerCAmelCase__ : Union[str, Any]=3 , lowerCAmelCase__ : Optional[Any]=None , ): SCREAMING_SNAKE_CASE_: List[str] = parent SCREAMING_SNAKE_CASE_: Optional[int] = batch_size SCREAMING_SNAKE_CASE_: Optional[int] = image_size SCREAMING_SNAKE_CASE_: Tuple = num_channels SCREAMING_SNAKE_CASE_: Dict = embeddings_size SCREAMING_SNAKE_CASE_: Optional[Any] = hidden_sizes SCREAMING_SNAKE_CASE_: Dict = depths SCREAMING_SNAKE_CASE_: Dict = is_training SCREAMING_SNAKE_CASE_: Any = use_labels SCREAMING_SNAKE_CASE_: Union[str, Any] = hidden_act SCREAMING_SNAKE_CASE_: Optional[int] = num_labels SCREAMING_SNAKE_CASE_: Optional[Any] = scope SCREAMING_SNAKE_CASE_: Optional[Any] = len(_a) def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) SCREAMING_SNAKE_CASE_: Optional[int] = None if self.use_labels: SCREAMING_SNAKE_CASE_: List[Any] = ids_tensor([self.batch_size] , self.num_labels) SCREAMING_SNAKE_CASE_: Union[str, Any] = self.get_config() return config, pixel_values, labels def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : str): SCREAMING_SNAKE_CASE_: List[Any] = TFResNetModel(config=_a) SCREAMING_SNAKE_CASE_: List[Any] = model(_a) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[str]): SCREAMING_SNAKE_CASE_: str = self.num_labels SCREAMING_SNAKE_CASE_: Tuple = TFResNetForImageClassification(_a) SCREAMING_SNAKE_CASE_: Tuple = model(_a , labels=_a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: List[Any] = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_: Dict = config_and_inputs SCREAMING_SNAKE_CASE_: Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class __lowercase ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : List[str] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () _UpperCAmelCase : Tuple = ( {'''feature-extraction''': TFResNetModel, '''image-classification''': TFResNetForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase : int = False _UpperCAmelCase : Union[str, Any] = False _UpperCAmelCase : List[Any] = False _UpperCAmelCase : List[Any] = False _UpperCAmelCase : List[str] = False def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_: Dict = TFResNetModelTester(self) SCREAMING_SNAKE_CASE_: Optional[Any] = ConfigTester(self , config_class=_a , has_text_modality=_a) def _SCREAMING_SNAKE_CASE ( self : int): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): return @unittest.skip(reason="ResNet does not use inputs_embeds") def _SCREAMING_SNAKE_CASE ( self : str): pass @unittest.skip(reason="ResNet does not support input and output embeddings") def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): pass def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_: List[Any] = model_class(_a) SCREAMING_SNAKE_CASE_: Any = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_: Tuple = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_: Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): def check_hidden_states_output(lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[Any]): SCREAMING_SNAKE_CASE_: int = model_class(_a) SCREAMING_SNAKE_CASE_: Any = model(**self._prepare_for_class(_a , _a)) SCREAMING_SNAKE_CASE_: Any = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states SCREAMING_SNAKE_CASE_: List[str] = self.model_tester.num_stages self.assertEqual(len(_a) , expected_num_stages + 1) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) SCREAMING_SNAKE_CASE_: Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_: Any = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: SCREAMING_SNAKE_CASE_: Union[str, Any] = layer_type SCREAMING_SNAKE_CASE_: str = True check_hidden_states_output(_a , _a , _a) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE_: Any = True check_hidden_states_output(_a , _a , _a) def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a) @slow def _SCREAMING_SNAKE_CASE ( self : List[Any]): for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_: Union[str, Any] = TFResNetModel.from_pretrained(_a) self.assertIsNotNone(_a) def A_ ( ): SCREAMING_SNAKE_CASE_: int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): """simple docstring""" @cached_property def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) SCREAMING_SNAKE_CASE_: str = self.default_image_processor SCREAMING_SNAKE_CASE_: List[str] = prepare_img() SCREAMING_SNAKE_CASE_: int = image_processor(images=_a , return_tensors="tf") # forward pass SCREAMING_SNAKE_CASE_: Tuple = model(**_a) # verify the logits SCREAMING_SNAKE_CASE_: Optional[Any] = tf.TensorShape((1, 1000)) self.assertEqual(outputs.logits.shape , _a) SCREAMING_SNAKE_CASE_: str = tf.constant([-11.1069, -9.7877, -8.3777]) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _a , atol=1E-4))
369
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer lowerCAmelCase : Any = logging.get_logger(__name__) lowerCAmelCase : Dict = {"""vocab_file""": """vocab.txt"""} lowerCAmelCase : List[str] = { """vocab_file""": { """YituTech/conv-bert-base""": """https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt""", """YituTech/conv-bert-medium-small""": ( """https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt""" ), """YituTech/conv-bert-small""": """https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt""", } } lowerCAmelCase : List[Any] = { """YituTech/conv-bert-base""": 512, """YituTech/conv-bert-medium-small""": 512, """YituTech/conv-bert-small""": 512, } lowerCAmelCase : Tuple = { """YituTech/conv-bert-base""": {"""do_lower_case""": True}, """YituTech/conv-bert-medium-small""": {"""do_lower_case""": True}, """YituTech/conv-bert-small""": {"""do_lower_case""": True}, } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : str = VOCAB_FILES_NAMES _UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : List[Any] = PRETRAINED_INIT_CONFIGURATION _UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Dict = ConvBertTokenizer def __init__( self : Optional[Any] , lowerCAmelCase__ : Optional[int]=None , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : str=True , lowerCAmelCase__ : Any="[UNK]" , lowerCAmelCase__ : Optional[Any]="[SEP]" , lowerCAmelCase__ : Any="[PAD]" , lowerCAmelCase__ : Dict="[CLS]" , lowerCAmelCase__ : Dict="[MASK]" , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : str=None , **lowerCAmelCase__ : Dict , ): super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , **lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get("lowercase" , lowerCAmelCase__) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase__) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase__) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE_: Optional[int] = getattr(lowerCAmelCase__ , normalizer_state.pop("type")) SCREAMING_SNAKE_CASE_: Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE_: List[str] = strip_accents SCREAMING_SNAKE_CASE_: Optional[Any] = tokenize_chinese_chars SCREAMING_SNAKE_CASE_: Optional[int] = normalizer_class(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = do_lower_case def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Union[str, Any]=None): SCREAMING_SNAKE_CASE_: List[str] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : List[int] , lowerCAmelCase__ : Optional[List[int]] = None): SCREAMING_SNAKE_CASE_: Optional[int] = [self.sep_token_id] SCREAMING_SNAKE_CASE_: int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[str] = None): SCREAMING_SNAKE_CASE_: Any = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__) return tuple(lowerCAmelCase__)
127
0
from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
13
def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = [0] * len(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] = [] SCREAMING_SNAKE_CASE_: str = [] SCREAMING_SNAKE_CASE_: List[str] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_UpperCAmelCase ) ): if indegree[i] == 0: queue.append(_UpperCAmelCase ) while queue: SCREAMING_SNAKE_CASE_: Optional[int] = queue.pop(0 ) cnt += 1 topo.append(_UpperCAmelCase ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(_UpperCAmelCase ) if cnt != len(_UpperCAmelCase ): print("Cycle exists" ) else: print(_UpperCAmelCase ) # Adjacency List of Graph lowerCAmelCase : Any = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
13
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool __UpperCAmelCase : List[str] = { "Acehnese Arabic": "ace_Arab", "Acehnese Latin": "ace_Latn", "Mesopotamian Arabic": "acm_Arab", "Ta\'izzi-Adeni Arabic": "acq_Arab", "Tunisian Arabic": "aeb_Arab", "Afrikaans": "afr_Latn", "South Levantine Arabic": "ajp_Arab", "Akan": "aka_Latn", "Amharic": "amh_Ethi", "North Levantine Arabic": "apc_Arab", "Modern Standard Arabic": "arb_Arab", "Modern Standard Arabic Romanized": "arb_Latn", "Najdi Arabic": "ars_Arab", "Moroccan Arabic": "ary_Arab", "Egyptian Arabic": "arz_Arab", "Assamese": "asm_Beng", "Asturian": "ast_Latn", "Awadhi": "awa_Deva", "Central Aymara": "ayr_Latn", "South Azerbaijani": "azb_Arab", "North Azerbaijani": "azj_Latn", "Bashkir": "bak_Cyrl", "Bambara": "bam_Latn", "Balinese": "ban_Latn", "Belarusian": "bel_Cyrl", "Bemba": "bem_Latn", "Bengali": "ben_Beng", "Bhojpuri": "bho_Deva", "Banjar Arabic": "bjn_Arab", "Banjar Latin": "bjn_Latn", "Standard Tibetan": "bod_Tibt", "Bosnian": "bos_Latn", "Buginese": "bug_Latn", "Bulgarian": "bul_Cyrl", "Catalan": "cat_Latn", "Cebuano": "ceb_Latn", "Czech": "ces_Latn", "Chokwe": "cjk_Latn", "Central Kurdish": "ckb_Arab", "Crimean Tatar": "crh_Latn", "Welsh": "cym_Latn", "Danish": "dan_Latn", "German": "deu_Latn", "Southwestern Dinka": "dik_Latn", "Dyula": "dyu_Latn", "Dzongkha": "dzo_Tibt", "Greek": "ell_Grek", "English": "eng_Latn", "Esperanto": "epo_Latn", "Estonian": "est_Latn", "Basque": "eus_Latn", "Ewe": "ewe_Latn", "Faroese": "fao_Latn", "Fijian": "fij_Latn", "Finnish": "fin_Latn", "Fon": "fon_Latn", "French": "fra_Latn", "Friulian": "fur_Latn", "Nigerian Fulfulde": "fuv_Latn", "Scottish Gaelic": "gla_Latn", "Irish": "gle_Latn", "Galician": "glg_Latn", "Guarani": "grn_Latn", "Gujarati": "guj_Gujr", "Haitian Creole": "hat_Latn", "Hausa": "hau_Latn", "Hebrew": "heb_Hebr", "Hindi": "hin_Deva", "Chhattisgarhi": "hne_Deva", "Croatian": "hrv_Latn", "Hungarian": "hun_Latn", "Armenian": "hye_Armn", "Igbo": "ibo_Latn", "Ilocano": "ilo_Latn", "Indonesian": "ind_Latn", "Icelandic": "isl_Latn", "Italian": "ita_Latn", "Javanese": "jav_Latn", "Japanese": "jpn_Jpan", "Kabyle": "kab_Latn", "Jingpho": "kac_Latn", "Kamba": "kam_Latn", "Kannada": "kan_Knda", "Kashmiri Arabic": "kas_Arab", "Kashmiri Devanagari": "kas_Deva", "Georgian": "kat_Geor", "Central Kanuri Arabic": "knc_Arab", "Central Kanuri Latin": "knc_Latn", "Kazakh": "kaz_Cyrl", "Kabiyè": "kbp_Latn", "Kabuverdianu": "kea_Latn", "Khmer": "khm_Khmr", "Kikuyu": "kik_Latn", "Kinyarwanda": "kin_Latn", "Kyrgyz": "kir_Cyrl", "Kimbundu": "kmb_Latn", "Northern Kurdish": "kmr_Latn", "Kikongo": "kon_Latn", "Korean": "kor_Hang", "Lao": "lao_Laoo", "Ligurian": "lij_Latn", "Limburgish": "lim_Latn", "Lingala": "lin_Latn", "Lithuanian": "lit_Latn", "Lombard": "lmo_Latn", "Latgalian": "ltg_Latn", "Luxembourgish": "ltz_Latn", "Luba-Kasai": "lua_Latn", "Ganda": "lug_Latn", "Luo": "luo_Latn", "Mizo": "lus_Latn", "Standard Latvian": "lvs_Latn", "Magahi": "mag_Deva", "Maithili": "mai_Deva", "Malayalam": "mal_Mlym", "Marathi": "mar_Deva", "Minangkabau Arabic ": "min_Arab", "Minangkabau Latin": "min_Latn", "Macedonian": "mkd_Cyrl", "Plateau Malagasy": "plt_Latn", "Maltese": "mlt_Latn", "Meitei Bengali": "mni_Beng", "Halh Mongolian": "khk_Cyrl", "Mossi": "mos_Latn", "Maori": "mri_Latn", "Burmese": "mya_Mymr", "Dutch": "nld_Latn", "Norwegian Nynorsk": "nno_Latn", "Norwegian Bokmål": "nob_Latn", "Nepali": "npi_Deva", "Northern Sotho": "nso_Latn", "Nuer": "nus_Latn", "Nyanja": "nya_Latn", "Occitan": "oci_Latn", "West Central Oromo": "gaz_Latn", "Odia": "ory_Orya", "Pangasinan": "pag_Latn", "Eastern Panjabi": "pan_Guru", "Papiamento": "pap_Latn", "Western Persian": "pes_Arab", "Polish": "pol_Latn", "Portuguese": "por_Latn", "Dari": "prs_Arab", "Southern Pashto": "pbt_Arab", "Ayacucho Quechua": "quy_Latn", "Romanian": "ron_Latn", "Rundi": "run_Latn", "Russian": "rus_Cyrl", "Sango": "sag_Latn", "Sanskrit": "san_Deva", "Santali": "sat_Olck", "Sicilian": "scn_Latn", "Shan": "shn_Mymr", "Sinhala": "sin_Sinh", "Slovak": "slk_Latn", "Slovenian": "slv_Latn", "Samoan": "smo_Latn", "Shona": "sna_Latn", "Sindhi": "snd_Arab", "Somali": "som_Latn", "Southern Sotho": "sot_Latn", "Spanish": "spa_Latn", "Tosk Albanian": "als_Latn", "Sardinian": "srd_Latn", "Serbian": "srp_Cyrl", "Swati": "ssw_Latn", "Sundanese": "sun_Latn", "Swedish": "swe_Latn", "Swahili": "swh_Latn", "Silesian": "szl_Latn", "Tamil": "tam_Taml", "Tatar": "tat_Cyrl", "Telugu": "tel_Telu", "Tajik": "tgk_Cyrl", "Tagalog": "tgl_Latn", "Thai": "tha_Thai", "Tigrinya": "tir_Ethi", "Tamasheq Latin": "taq_Latn", "Tamasheq Tifinagh": "taq_Tfng", "Tok Pisin": "tpi_Latn", "Tswana": "tsn_Latn", "Tsonga": "tso_Latn", "Turkmen": "tuk_Latn", "Tumbuka": "tum_Latn", "Turkish": "tur_Latn", "Twi": "twi_Latn", "Central Atlas Tamazight": "tzm_Tfng", "Uyghur": "uig_Arab", "Ukrainian": "ukr_Cyrl", "Umbundu": "umb_Latn", "Urdu": "urd_Arab", "Northern Uzbek": "uzn_Latn", "Venetian": "vec_Latn", "Vietnamese": "vie_Latn", "Waray": "war_Latn", "Wolof": "wol_Latn", "Xhosa": "xho_Latn", "Eastern Yiddish": "ydd_Hebr", "Yoruba": "yor_Latn", "Yue Chinese": "yue_Hant", "Chinese Simplified": "zho_Hans", "Chinese Traditional": "zho_Hant", "Standard Malay": "zsm_Latn", "Zulu": "zul_Latn", } class __snake_case ( snake_case_ ): '''simple docstring''' lowerCAmelCase__ = """facebook/nllb-200-distilled-600M""" lowerCAmelCase__ = ( """This is a tool that translates text from a language to another. It takes three inputs: `text`, which should """ """be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, """ """which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in """ """plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.""" ) lowerCAmelCase__ = """translator""" lowerCAmelCase__ = AutoTokenizer lowerCAmelCase__ = AutoModelForSeqaSeqLM lowerCAmelCase__ = LANGUAGE_CODES lowerCAmelCase__ = ["""text""", """text""", """text"""] lowerCAmelCase__ = ["""text"""] def UpperCAmelCase__ ( self : int , A : Tuple , A : Union[str, Any] , A : List[Any] ): if src_lang not in self.lang_to_code: raise ValueError(f'''{src_lang} is not a supported language.''' ) if tgt_lang not in self.lang_to_code: raise ValueError(f'''{tgt_lang} is not a supported language.''' ) __snake_case: List[str] = self.lang_to_code[src_lang] __snake_case: List[str] = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( A , return_tensors="""pt""" , src_lang=A , tgt_lang=A ) def UpperCAmelCase__ ( self : Any , A : Tuple ): return self.model.generate(**A ) def UpperCAmelCase__ ( self : Tuple , A : List[str] ): return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=A )
353
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __UpperCAmelCase : str = logging.get_logger(__name__) class __snake_case ( __lowerCamelCase ): '''simple docstring''' def __init__( self : Any , A : int , A : int , A : float , **A : Optional[int] ): __snake_case: List[str] = feature_size __snake_case: Optional[int] = sampling_rate __snake_case: Any = padding_value __snake_case: Dict = kwargs.pop("""padding_side""" , """right""" ) __snake_case: Union[str, Any] = kwargs.pop("""return_attention_mask""" , A ) super().__init__(**A ) def UpperCAmelCase__ ( self : Optional[Any] , A : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , A : Union[bool, str, PaddingStrategy] = True , A : Optional[int] = None , A : bool = False , A : Optional[int] = None , A : Optional[bool] = None , A : Optional[Union[str, TensorType]] = None , ): # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): __snake_case: Optional[int] = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( """You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`""" f''' to this method that includes {self.model_input_names[0]}, but you provided''' f''' {list(processed_features.keys() )}''' ) __snake_case: List[str] = processed_features[self.model_input_names[0]] __snake_case: Any = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A ) == 0: if return_attention_mask: __snake_case: Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch __snake_case: int = required_input[0] if isinstance(A , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. __snake_case: Optional[int] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A ): __snake_case: Optional[int] = required_input[index][0] if return_tensors is None: if is_tf_tensor(A ): __snake_case: str = """tf""" elif is_torch_tensor(A ): __snake_case: str = """pt""" elif isinstance(A , (int, float, list, tuple, np.ndarray) ): __snake_case: List[str] = """np""" else: raise ValueError( f'''type of {first_element} unknown: {type(A )}. ''' """Should be one of a python, numpy, pytorch or tensorflow object.""" ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): __snake_case: List[Any] = to_numpy(A ) else: __snake_case: Union[str, Any] = [to_numpy(A ) for v in value] # Convert padding_strategy in PaddingStrategy __snake_case: Union[str, Any] = self._get_padding_strategies(padding=A , max_length=A ) __snake_case: Any = processed_features[self.model_input_names[0]] __snake_case: int = len(A ) if not all(len(A ) == batch_size for v in processed_features.values() ): raise ValueError("""Some items in the output dictionary have a different batch size than others.""" ) __snake_case: Union[str, Any] = [] for i in range(A ): __snake_case: List[Any] = {k: v[i] for k, v in processed_features.items()} # truncation __snake_case: Tuple = self._truncate( A , max_length=A , pad_to_multiple_of=A , truncation=A , ) truncated_inputs.append(A ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length __snake_case: Optional[Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) __snake_case: List[str] = PaddingStrategy.MAX_LENGTH __snake_case: List[Any] = {} for i in range(A ): # padding __snake_case: Any = self._pad( truncated_inputs[i] , max_length=A , padding_strategy=A , pad_to_multiple_of=A , return_attention_mask=A , ) for key, value in outputs.items(): if key not in batch_outputs: __snake_case: Optional[Any] = [] if value.dtype is np.dtype(np.floataa ): __snake_case: str = value.astype(np.floataa ) batch_outputs[key].append(A ) return BatchFeature(A , tensor_type=A ) def UpperCAmelCase__ ( self : int , A : Union[Dict[str, np.ndarray], BatchFeature] , A : Optional[int] = None , A : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , A : Optional[int] = None , A : Optional[bool] = None , ): __snake_case: List[Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: __snake_case: List[str] = len(A ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): __snake_case: List[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of __snake_case: Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A ) < max_length if return_attention_mask and "attention_mask" not in processed_features: __snake_case: List[str] = np.ones(len(A ) , dtype=np.intaa ) if needs_to_be_padded: __snake_case: Any = max_length - len(A ) if self.padding_side == "right": if return_attention_mask: __snake_case: Optional[int] = np.pad( processed_features["""attention_mask"""] , (0, difference) ) __snake_case: Any = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) __snake_case: Union[str, Any] = np.pad( A , A , """constant""" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: __snake_case: Dict = np.pad( processed_features["""attention_mask"""] , (difference, 0) ) __snake_case: Union[str, Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) __snake_case: str = np.pad( A , A , """constant""" , constant_values=self.padding_value ) else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return processed_features def UpperCAmelCase__ ( self : Optional[Any] , A : Union[Dict[str, np.ndarray], BatchFeature] , A : Optional[int] = None , A : Optional[int] = None , A : Optional[bool] = None , ): if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("""When setting ``truncation=True``, make sure that ``max_length`` is defined.""" ) __snake_case: List[str] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): __snake_case: List[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of __snake_case: Tuple = len(A ) > max_length if needs_to_be_truncated: __snake_case: List[Any] = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: __snake_case: int = processed_features["""attention_mask"""][:max_length] return processed_features def UpperCAmelCase__ ( self : int , A : int=False , A : int=None ): # Get padding strategy if padding is not False: if padding is True: __snake_case: Optional[int] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A , A ): __snake_case: Optional[int] = PaddingStrategy(A ) elif isinstance(A , A ): __snake_case: Any = padding else: __snake_case: Any = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( f'''When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined''' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( """Asking to pad but the feature_extractor does not have a padding value. Please select a value to use""" """ as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.""" ) return padding_strategy
293
0
def a__ ( A_, A_ ): '''simple docstring''' def get_matched_characters(A_, A_ ) -> str: __magic_name__ = [] __magic_name__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): __magic_name__ = int(max(0, i - limit ) ) __magic_name__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(A_ ) __magic_name__ = f'''{_stra[0:_stra.index(A_ )]} {_stra[_stra.index(A_ ) + 1:]}''' return "".join(A_ ) # matching characters __magic_name__ = get_matched_characters(A_, A_ ) __magic_name__ = get_matched_characters(A_, A_ ) __magic_name__ = len(A_ ) # transposition __magic_name__ = ( len([(ca, ca) for ca, ca in zip(A_, A_ ) if ca != ca] ) // 2 ) if not match_count: __magic_name__ = 0.0 else: __magic_name__ = ( 1 / 3 * ( match_count / len(A_ ) + match_count / len(A_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters __magic_name__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
88
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowercase ( self : List[str] , UpperCamelCase__ : int ) -> str: """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): __magic_name__ = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCamelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __magic_name__ = """sgugger/tiny-distilbert-classification""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , only_pretrain_model=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" __magic_name__ = """patrickvonplaten/t5-tiny-random""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , configs=[config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , """Cannot do xla on CPU.""" ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , use_xla=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Union[str, Any] ) -> Dict: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , save_to_csv=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCamelCase__ , """inf_time.csv""" ) , inference_memory_csv_file=os.path.join(UpperCamelCase__ , """inf_mem.csv""" ) , env_info_csv_file=os.path.join(UpperCamelCase__ , """env.csv""" ) , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCamelCase__ , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """env.csv""" ) ).exists() ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCamelCase__ : Dict ): self.assertTrue(hasattr(UpperCamelCase__ , """sequential""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """cumulative""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """current""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCamelCase__ , """log.txt""" ) , log_print=UpperCamelCase__ , trace_memory_line_by_line=UpperCamelCase__ , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """log.txt""" ) ).exists() )
88
1
'''simple docstring''' import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class snake_case__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" lowerCamelCase = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def lowerCAmelCase ( self : str , UpperCamelCase__ : List[str]=0 ) -> Tuple: """simple docstring""" snake_case : List[Any] = np.random.RandomState(UpperCamelCase__ ) snake_case : List[str] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCAmelCase ( self : Any ) -> Optional[Any]: """simple docstring""" snake_case : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : List[str] = self.get_dummy_inputs() snake_case : Optional[int] = pipe(**UpperCamelCase__ ).images snake_case : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : Union[str, Any] = np.array([0.65_072, 0.58_492, 0.48_219, 0.55_521, 0.53_180, 0.55_939, 0.50_697, 0.39_800, 0.46_455] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : str ) -> List[str]: """simple docstring""" snake_case : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) snake_case : Optional[int] = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Optional[Any] = self.get_dummy_inputs() snake_case : List[Any] = pipe(**UpperCamelCase__ ).images snake_case : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : Optional[Any] = np.array([0.65_863, 0.59_425, 0.49_326, 0.56_313, 0.53_875, 0.56_627, 0.51_065, 0.39_777, 0.46_330] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : Tuple ) -> int: """simple docstring""" snake_case : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) snake_case : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Tuple = self.get_dummy_inputs() snake_case : str = pipe(**UpperCamelCase__ ).images snake_case : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : str = np.array([0.53_755, 0.60_786, 0.47_402, 0.49_488, 0.51_869, 0.49_819, 0.47_985, 0.38_957, 0.44_279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : Dict ) -> List[str]: """simple docstring""" snake_case : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) snake_case : Optional[int] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Tuple = self.get_dummy_inputs() snake_case : Union[str, Any] = pipe(**UpperCamelCase__ ).images snake_case : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : Union[str, Any] = np.array([0.53_755, 0.60_786, 0.47_402, 0.49_488, 0.51_869, 0.49_819, 0.47_985, 0.38_957, 0.44_279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" snake_case : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) snake_case : str = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Optional[int] = self.get_dummy_inputs() snake_case : Any = pipe(**UpperCamelCase__ ).images snake_case : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : Tuple = np.array([0.53_817, 0.60_812, 0.47_384, 0.49_530, 0.51_894, 0.49_814, 0.47_984, 0.38_958, 0.44_271] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : str ) -> int: """simple docstring""" snake_case : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) snake_case : List[str] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Tuple = self.get_dummy_inputs() snake_case : int = pipe(**UpperCamelCase__ ).images snake_case : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case : Dict = np.array([0.53_895, 0.60_808, 0.47_933, 0.49_608, 0.51_886, 0.49_950, 0.48_053, 0.38_957, 0.44_200] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase ( self : int ) -> Optional[int]: """simple docstring""" snake_case : Any = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : List[str] = self.get_dummy_inputs() snake_case : Any = 3 * [inputs['''prompt''']] # forward snake_case : List[str] = pipe(**UpperCamelCase__ ) snake_case : Tuple = output.images[0, -3:, -3:, -1] snake_case : Tuple = self.get_dummy_inputs() snake_case : List[Any] = 3 * [inputs.pop('''prompt''' )] snake_case : List[str] = pipe.tokenizer( UpperCamelCase__ , padding='''max_length''' , max_length=pipe.tokenizer.model_max_length , truncation=UpperCamelCase__ , return_tensors='''np''' , ) snake_case : Tuple = text_inputs['''input_ids'''] snake_case : int = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] snake_case : List[Any] = prompt_embeds # forward snake_case : Optional[Any] = pipe(**UpperCamelCase__ ) snake_case : List[Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 def lowerCAmelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" snake_case : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Optional[int] = self.get_dummy_inputs() snake_case : Any = 3 * ['''this is a negative prompt'''] snake_case : Any = negative_prompt snake_case : Tuple = 3 * [inputs['''prompt''']] # forward snake_case : List[str] = pipe(**UpperCamelCase__ ) snake_case : Any = output.images[0, -3:, -3:, -1] snake_case : Tuple = self.get_dummy_inputs() snake_case : Optional[Any] = 3 * [inputs.pop('''prompt''' )] snake_case : str = [] for p in [prompt, negative_prompt]: snake_case : int = pipe.tokenizer( UpperCamelCase__ , padding='''max_length''' , max_length=pipe.tokenizer.model_max_length , truncation=UpperCamelCase__ , return_tensors='''np''' , ) snake_case : int = text_inputs['''input_ids'''] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) snake_case ,snake_case : List[str] = embeds # forward snake_case : List[Any] = pipe(**UpperCamelCase__ ) snake_case : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @nightly @require_onnxruntime @require_torch_gpu class snake_case__ ( unittest.TestCase ): """simple docstring""" @property def lowerCAmelCase ( self : int ) -> Dict: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCAmelCase ( self : List[str] ) -> Any: """simple docstring""" snake_case : Any = ort.SessionOptions() snake_case : Optional[Any] = False return options def lowerCAmelCase ( self : Any ) -> str: """simple docstring""" snake_case : Tuple = OnnxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Union[str, Any] = '''A painting of a squirrel eating a burger''' np.random.seed(0 ) snake_case : List[str] = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type='''np''' ) snake_case : Union[str, Any] = output.images snake_case : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case : Tuple = np.array([0.0_452, 0.0_390, 0.0_087, 0.0_350, 0.0_617, 0.0_364, 0.0_544, 0.0_523, 0.0_720] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase ( self : Optional[Any] ) -> Any: """simple docstring""" snake_case : Optional[Any] = DDIMScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) snake_case : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : int = '''open neural network exchange''' snake_case : int = np.random.RandomState(0 ) snake_case : str = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type='''np''' ) snake_case : Optional[Any] = output.images snake_case : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case : str = np.array([0.2_867, 0.1_974, 0.1_481, 0.7_294, 0.7_251, 0.6_667, 0.4_194, 0.5_642, 0.6_486] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" snake_case : Tuple = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) snake_case : str = OnnxStableDiffusionPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : List[str] = '''open neural network exchange''' snake_case : Tuple = np.random.RandomState(0 ) snake_case : int = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type='''np''' ) snake_case : List[Any] = output.images snake_case : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case : str = np.array([0.2_306, 0.1_959, 0.1_593, 0.6_549, 0.6_394, 0.5_408, 0.5_065, 0.6_010, 0.6_161] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase ( self : List[str] ) -> Dict: """simple docstring""" snake_case : Tuple = 0 def test_callback_fn(UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : np.ndarray ) -> None: snake_case : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) snake_case : str = latents[0, -3:, -3:, -1] snake_case : Any = np.array( [-0.6_772, -0.3_835, -1.2_456, 0.1_905, -1.0_974, 0.6_967, -1.9_353, 0.0_178, 1.0_167] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) snake_case : List[Any] = latents[0, -3:, -3:, -1] snake_case : int = np.array( [-0.3_351, 0.2_241, -0.1_837, -0.2_325, -0.6_577, 0.3_393, -0.0_241, 0.5_899, 1.3_875] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 snake_case : List[str] = False snake_case : Any = OnnxStableDiffusionPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Tuple = '''Andromeda galaxy in a bottle''' snake_case : Union[str, Any] = np.random.RandomState(0 ) pipe( prompt=UpperCamelCase__ , num_inference_steps=5 , guidance_scale=7.5 , generator=UpperCamelCase__ , callback=UpperCamelCase__ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def lowerCAmelCase ( self : Any ) -> List[str]: """simple docstring""" snake_case : str = OnnxStableDiffusionPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) assert pipe.safety_checker is None snake_case : List[Any] = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCamelCase__ ) snake_case : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(UpperCamelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None snake_case : Tuple = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None
83
'''simple docstring''' import os def _UpperCamelCase ( ) -> List[Any]: '''simple docstring''' snake_case : int = os.path.join(os.path.dirname(SCREAMING_SNAKE_CASE__ ) , '''num.txt''' ) with open(SCREAMING_SNAKE_CASE__ ) as file_hand: return str(sum(int(SCREAMING_SNAKE_CASE__ ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
83
1
"""simple docstring""" from math import isclose, sqrt def _snake_case ( lowercase__ : float , lowercase__ : float , lowercase__ : float ) -> tuple[float, float, float]: '''simple docstring''' lowerCAmelCase_ :Optional[int] = point_y / 4 / point_x lowerCAmelCase_ :Dict = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) lowerCAmelCase_ :Union[str, Any] = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) lowerCAmelCase_ :str = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 lowerCAmelCase_ :Tuple = outgoing_gradient**2 + 4 lowerCAmelCase_ :Tuple = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) lowerCAmelCase_ :str = (point_y - outgoing_gradient * point_x) ** 2 - 1_0_0 lowerCAmelCase_ :Optional[Any] = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) lowerCAmelCase_ :Optional[int] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point lowerCAmelCase_ :List[Any] = x_minus if isclose(lowercase__ , lowercase__ ) else x_plus lowerCAmelCase_ :List[str] = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def _snake_case ( lowercase__ : float = 1.4 , lowercase__ : float = -9.6 ) -> int: '''simple docstring''' lowerCAmelCase_ :int = 0 lowerCAmelCase_ :float = first_x_coord lowerCAmelCase_ :float = first_y_coord lowerCAmelCase_ :float = (10.1 - point_y) / (0.0 - point_x) while not (-0.01 <= point_x <= 0.01 and point_y > 0): lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ :List[str] = next_point(lowercase__ , lowercase__ , lowercase__ ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"""{solution() = }""")
84
"""simple docstring""" import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _SCREAMING_SNAKE_CASE ( A__ , unittest.TestCase ): UpperCAmelCase_ :Any = BioGptTokenizer UpperCAmelCase_ :str = False def __lowerCAmelCase ( self ) -> List[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase_ :Optional[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] lowerCAmelCase_ :str = dict(zip(__A , range(len(__A ) ) ) ) lowerCAmelCase_ :int = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] lowerCAmelCase_ :Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCAmelCase_ :Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def __lowerCAmelCase ( self , __A ) -> Optional[int]: lowerCAmelCase_ :List[Any] = """lower newer""" lowerCAmelCase_ :Tuple = """lower newer""" return input_text, output_text def __lowerCAmelCase ( self ) -> str: lowerCAmelCase_ :List[str] = BioGptTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase_ :Union[str, Any] = """lower""" lowerCAmelCase_ :Any = ["""low""", """er</w>"""] lowerCAmelCase_ :Union[str, Any] = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) lowerCAmelCase_ :Dict = tokens + ["""<unk>"""] lowerCAmelCase_ :List[str] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) @slow def __lowerCAmelCase ( self ) -> List[Any]: lowerCAmelCase_ :Optional[Any] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) lowerCAmelCase_ :List[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=__A ) lowerCAmelCase_ :List[str] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__A ) lowerCAmelCase_ :Optional[int] = tokenizer.build_inputs_with_special_tokens(__A ) lowerCAmelCase_ :List[str] = tokenizer.build_inputs_with_special_tokens(__A , __A ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
84
1
'''simple docstring''' import string import numpy def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return b if a == 0 else greatest_common_divisor(b % a , _snake_case ) class _lowercase : a = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) a = numpy.vectorize(lambda _lowercase : x % 36 ) a = numpy.vectorize(_lowercase ) def __init__( self: int , UpperCamelCase__: int ): lowerCamelCase__ : List[str] = self.modulus(_a ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key lowerCamelCase__ : Tuple = encrypt_key.shape[0] def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Tuple ): return self.key_string.index(_a ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Any ): return self.key_string[round(_a )] def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: lowerCamelCase__ : Tuple = det % len(self.key_string ) lowerCamelCase__ : List[str] = len(self.key_string ) if greatest_common_divisor(_a , len(self.key_string ) ) != 1: lowerCamelCase__ : List[str] = ( F'''determinant modular {req_l} of encryption key({det}) ''' F'''is not co prime w.r.t {req_l}.\nTry another key.''' ) raise ValueError(_a ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Dict = [char for char in text.upper() if char in self.key_string] lowerCamelCase__ : List[Any] = chars[-1] while len(_a ) % self.break_key != 0: chars.append(_a ) return "".join(_a ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int ): lowerCamelCase__ : Tuple = self.process_text(text.upper() ) lowerCamelCase__ : Union[str, Any] = "" for i in range(0 , len(_a ) - self.break_key + 1 , self.break_key ): lowerCamelCase__ : str = text[i : i + self.break_key] lowerCamelCase__ : Optional[Any] = [self.replace_letters(_a ) for char in batch] lowerCamelCase__ : List[str] = numpy.array([vec] ).T lowerCamelCase__ : List[str] = self.modulus(self.encrypt_key.dot(_a ) ).T.tolist()[ 0 ] lowerCamelCase__ : Optional[int] = "".join( self.replace_digits(_a ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Union[str, Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: lowerCamelCase__ : int = det % len(self.key_string ) lowerCamelCase__ : Dict = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: lowerCamelCase__ : str = i break lowerCamelCase__ : Union[str, Any] = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(_a ) ) def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Tuple = self.make_decrypt_key() lowerCamelCase__ : Dict = self.process_text(text.upper() ) lowerCamelCase__ : Any = "" for i in range(0 , len(_a ) - self.break_key + 1 , self.break_key ): lowerCamelCase__ : Tuple = text[i : i + self.break_key] lowerCamelCase__ : Any = [self.replace_letters(_a ) for char in batch] lowerCamelCase__ : Any = numpy.array([vec] ).T lowerCamelCase__ : List[Any] = self.modulus(decrypt_key.dot(_a ) ).T.tolist()[0] lowerCamelCase__ : Tuple = "".join( self.replace_digits(_a ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def SCREAMING_SNAKE_CASE_ () -> None: lowerCamelCase__ : Union[str, Any] = int(input("""Enter the order of the encryption key: """ ) ) lowerCamelCase__ : str = [] print("""Enter each row of the encryption key with space separated integers""" ) for _ in range(_snake_case ): lowerCamelCase__ : str = [int(_snake_case ) for x in input().split()] hill_matrix.append(_snake_case ) lowerCamelCase__ : Any = HillCipher(numpy.array(_snake_case ) ) print("""Would you like to encrypt or decrypt some text? (1 or 2)""" ) lowerCamelCase__ : List[Any] = input("""\n1. Encrypt\n2. Decrypt\n""" ) if option == "1": lowerCamelCase__ : str = input("""What text would you like to encrypt?: """ ) print("""Your encrypted text is:""" ) print(hc.encrypt(_snake_case ) ) elif option == "2": lowerCamelCase__ : Any = input("""What text would you like to decrypt?: """ ) print("""Your decrypted text is:""" ) print(hc.decrypt(_snake_case ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
369
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _A : Optional[Any] =logging.get_logger(__name__) _A : List[str] ={ '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class _lowercase ( _lowercase ): a = """marian""" a = ["""past_key_values"""] a = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self: Tuple , UpperCamelCase__: Optional[Any]=58_101 , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: Union[str, Any]=1_024 , UpperCamelCase__: Any=12 , UpperCamelCase__: Optional[int]=4_096 , UpperCamelCase__: Tuple=16 , UpperCamelCase__: Dict=12 , UpperCamelCase__: Optional[Any]=4_096 , UpperCamelCase__: Any=16 , UpperCamelCase__: List[str]=0.0 , UpperCamelCase__: Tuple=0.0 , UpperCamelCase__: str=True , UpperCamelCase__: Optional[int]=True , UpperCamelCase__: Optional[int]="gelu" , UpperCamelCase__: Union[str, Any]=1_024 , UpperCamelCase__: Optional[int]=0.1 , UpperCamelCase__: Optional[Any]=0.0 , UpperCamelCase__: Optional[Any]=0.0 , UpperCamelCase__: Optional[int]=0.02 , UpperCamelCase__: str=58_100 , UpperCamelCase__: Tuple=False , UpperCamelCase__: Optional[Any]=58_100 , UpperCamelCase__: int=0 , UpperCamelCase__: Union[str, Any]=0 , UpperCamelCase__: List[str]=True , **UpperCamelCase__: str , ): lowerCamelCase__ : int = vocab_size lowerCamelCase__ : Tuple = decoder_vocab_size or vocab_size lowerCamelCase__ : List[str] = max_position_embeddings lowerCamelCase__ : Optional[Any] = d_model lowerCamelCase__ : int = encoder_ffn_dim lowerCamelCase__ : Union[str, Any] = encoder_layers lowerCamelCase__ : Dict = encoder_attention_heads lowerCamelCase__ : Optional[int] = decoder_ffn_dim lowerCamelCase__ : List[str] = decoder_layers lowerCamelCase__ : Dict = decoder_attention_heads lowerCamelCase__ : int = dropout lowerCamelCase__ : str = attention_dropout lowerCamelCase__ : Dict = activation_dropout lowerCamelCase__ : List[str] = activation_function lowerCamelCase__ : Union[str, Any] = init_std lowerCamelCase__ : str = encoder_layerdrop lowerCamelCase__ : Any = decoder_layerdrop lowerCamelCase__ : List[str] = use_cache lowerCamelCase__ : List[str] = encoder_layers lowerCamelCase__ : int = scale_embedding # scale factor will be sqrt(d_model) if True lowerCamelCase__ : str = share_encoder_decoder_embeddings super().__init__( pad_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , is_encoder_decoder=UpperCamelCase__ , decoder_start_token_id=UpperCamelCase__ , forced_eos_token_id=UpperCamelCase__ , **UpperCamelCase__ , ) class _lowercase ( _lowercase ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def lowerCamelCase_ ( self: Union[str, Any] ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : List[str] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: lowerCamelCase__ : Dict = {0: """batch"""} lowerCamelCase__ : Union[str, Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: lowerCamelCase__ : Any = {0: """batch""", 1: """decoder_sequence"""} lowerCamelCase__ : Dict = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(UpperCamelCase__ , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCamelCase__ : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: lowerCamelCase__ , lowerCamelCase__ : Tuple = self.num_layers for i in range(UpperCamelCase__ ): lowerCamelCase__ : Union[str, Any] = {0: """batch""", 2: """past_sequence + sequence"""} lowerCamelCase__ : List[str] = {0: """batch""", 2: """past_sequence + sequence"""} else: lowerCamelCase__ : int = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def lowerCamelCase_ ( self: Optional[Any] ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Dict = super().outputs else: lowerCamelCase__ : Any = super(UpperCamelCase__ , self ).outputs if self.use_past: lowerCamelCase__ , lowerCamelCase__ : str = self.num_layers for i in range(UpperCamelCase__ ): lowerCamelCase__ : Tuple = {0: """batch""", 2: """past_sequence + sequence"""} lowerCamelCase__ : Union[str, Any] = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def lowerCamelCase_ ( self: str , UpperCamelCase__: PreTrainedTokenizer , UpperCamelCase__: int = -1 , UpperCamelCase__: int = -1 , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[TensorType] = None , ): lowerCamelCase__ : Union[str, Any] = self._generate_dummy_inputs_for_encoder_and_decoder( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Generate decoder inputs lowerCamelCase__ : Any = seq_length if not self.use_past else 1 lowerCamelCase__ : Optional[Any] = self._generate_dummy_inputs_for_encoder_and_decoder( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} lowerCamelCase__ : Optional[int] = dict(**UpperCamelCase__ , **UpperCamelCase__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = common_inputs["""input_ids"""].shape lowerCamelCase__ : Tuple = common_inputs["""decoder_input_ids"""].shape[1] lowerCamelCase__ , lowerCamelCase__ : List[str] = self.num_attention_heads lowerCamelCase__ : Any = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Tuple = decoder_seq_length + 3 lowerCamelCase__ : int = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCamelCase__ : Optional[int] = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(UpperCamelCase__ , UpperCamelCase__ )] , dim=1 ) lowerCamelCase__ : Any = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCamelCase__ , lowerCamelCase__ : Any = self.num_layers lowerCamelCase__ : str = min(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = max(UpperCamelCase__ , UpperCamelCase__ ) - min_num_layers lowerCamelCase__ : int = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(UpperCamelCase__ ): common_inputs["past_key_values"].append( ( torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ ), ) ) # TODO: test this. lowerCamelCase__ : Union[str, Any] = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(UpperCamelCase__ , UpperCamelCase__ ): common_inputs["past_key_values"].append((torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) ) return common_inputs def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: PreTrainedTokenizer , UpperCamelCase__: int = -1 , UpperCamelCase__: int = -1 , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[TensorType] = None , ): lowerCamelCase__ : Any = self._generate_dummy_inputs_for_encoder_and_decoder( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase__ , lowerCamelCase__ : Any = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase__ : Optional[Any] = seqlen + 2 lowerCamelCase__ , lowerCamelCase__ : Dict = self.num_layers lowerCamelCase__ , lowerCamelCase__ : Dict = self.num_attention_heads lowerCamelCase__ : Optional[Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Optional[Any] = common_inputs["""attention_mask"""].dtype lowerCamelCase__ : int = torch.cat( [common_inputs["""attention_mask"""], torch.ones(UpperCamelCase__ , UpperCamelCase__ , dtype=UpperCamelCase__ )] , dim=1 ) lowerCamelCase__ : int = [ (torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) for _ in range(UpperCamelCase__ ) ] return common_inputs def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: PreTrainedTokenizer , UpperCamelCase__: int = -1 , UpperCamelCase__: int = -1 , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[TensorType] = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowerCamelCase__ : List[Any] = compute_effective_axis_dimension( UpperCamelCase__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCamelCase__ : Union[str, Any] = tokenizer.num_special_tokens_to_add(UpperCamelCase__ ) lowerCamelCase__ : Any = compute_effective_axis_dimension( UpperCamelCase__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=UpperCamelCase__ ) # Generate dummy inputs according to compute batch and sequence lowerCamelCase__ : Union[str, Any] = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCamelCase__ : str = dict(tokenizer(UpperCamelCase__ , return_tensors=UpperCamelCase__ ) ) return common_inputs def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: PreTrainedTokenizer , UpperCamelCase__: int = -1 , UpperCamelCase__: int = -1 , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[TensorType] = None , ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Dict = self._generate_dummy_inputs_for_default_and_seqaseq_lm( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) else: lowerCamelCase__ : Tuple = self._generate_dummy_inputs_for_causal_lm( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) return common_inputs def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: Optional[Any] ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Dict = super()._flatten_past_key_values_(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else: lowerCamelCase__ : List[Any] = super(UpperCamelCase__ , self )._flatten_past_key_values_( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) @property def lowerCamelCase_ ( self: Union[str, Any] ): return 1e-4
129
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCamelCase : List[str] = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class a ( _lowercase ): UpperCAmelCase_ : Dict ="distilbert" UpperCAmelCase_ : Optional[Any] ={ "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , _lowerCamelCase=3_0_5_2_2 , _lowerCamelCase=5_1_2 , _lowerCamelCase=False , _lowerCamelCase=6 , _lowerCamelCase=1_2 , _lowerCamelCase=7_6_8 , _lowerCamelCase=4 * 7_6_8 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase="gelu" , _lowerCamelCase=0.0_2 , _lowerCamelCase=0.1 , _lowerCamelCase=0.2 , _lowerCamelCase=0 , **_lowerCamelCase , ): lowercase = vocab_size lowercase = max_position_embeddings lowercase = sinusoidal_pos_embds lowercase = n_layers lowercase = n_heads lowercase = dim lowercase = hidden_dim lowercase = dropout lowercase = attention_dropout lowercase = activation lowercase = initializer_range lowercase = qa_dropout lowercase = seq_classif_dropout super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase ) class a ( _lowercase ): @property def UpperCamelCase_ ( self ): if self.task == "multiple-choice": lowercase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
220
'''simple docstring''' import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def lowercase_ ( _lowercase ) -> List[Any]: '''simple docstring''' if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class __lowercase ( nn.Module ): def __init__(self , A , A ): super().__init__() lowerCamelCase_ : Tuple = module lowerCamelCase_ : Any = nn.Sequential( nn.Linear(module.in_features , A , bias=A ) , nn.Linear(A , module.out_features , bias=A ) , ) lowerCamelCase_ : Optional[Any] = (2.0 / (5 * min(module.in_features , module.out_features ))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=A ) nn.init.zeros_(self.adapter[1].weight ) self.adapter.to(module.weight.device ) def UpperCAmelCase__ (self , A , *A , **A ): return self.module(A , *A , **A ) + self.adapter(A ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __lowercase ( unittest.TestCase ): # We keep the constants inside the init function and model loading inside setUp function # We need to test on relatively large models (aka >1b parameters otherwise the quantiztion may not work as expected) # Therefore here we use only bloom-1b3 to test our module lowerCamelCase : Tuple = "bigscience/bloom-1b7" # Constant values lowerCamelCase : List[Any] = 2.1_0_9_6_5_9_5_5_2_6_9_2_5_7_4 lowerCamelCase : int = "Hello my name is" lowerCamelCase : Tuple = set() EXPECTED_OUTPUTS.add("Hello my name is John and I am a professional photographer. I" ) EXPECTED_OUTPUTS.add("Hello my name is John.\nI am a friend of your father.\n" ) EXPECTED_OUTPUTS.add("Hello my name is John Doe, I am a student at the University" ) lowerCamelCase : Optional[int] = 10 def UpperCAmelCase__ (self ): # Models and tokenizer lowerCamelCase_ : Optional[Any] = AutoTokenizer.from_pretrained(self.model_name ) class __lowercase ( _lowercase ): def UpperCAmelCase__ (self ): super().setUp() # Models and tokenizer lowerCamelCase_ : Tuple = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map='''auto''' ) lowerCamelCase_ : List[str] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=A , device_map='''auto''' ) def UpperCAmelCase__ (self ): del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ (self ): lowerCamelCase_ : Dict = self.model_abit.config self.assertTrue(hasattr(A , '''quantization_config''' ) ) lowerCamelCase_ : Tuple = config.to_dict() lowerCamelCase_ : Optional[Any] = config.to_diff_dict() lowerCamelCase_ : Any = config.to_json_string() def UpperCAmelCase__ (self ): from bitsandbytes.nn import Paramsabit lowerCamelCase_ : str = self.model_fpaa.get_memory_footprint() lowerCamelCase_ : List[str] = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE ) lowerCamelCase_ : Optional[int] = get_some_linear_layer(self.model_abit ) self.assertTrue(linear.weight.__class__ == Paramsabit ) def UpperCAmelCase__ (self ): from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(A , torch.nn.Linear ): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta ) def UpperCAmelCase__ (self ): lowerCamelCase_ : Union[str, Any] = self.tokenizer(self.input_text , return_tensors='''pt''' ) lowerCamelCase_ : Union[str, Any] = self.model_abit.generate(input_ids=encoded_input['''input_ids'''].to(0 ) , max_new_tokens=1_0 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=A ) , self.EXPECTED_OUTPUTS ) def UpperCAmelCase__ (self ): lowerCamelCase_ : Dict = BitsAndBytesConfig() lowerCamelCase_ : Optional[Any] = True lowerCamelCase_ : str = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=A , device_map='''auto''' ) lowerCamelCase_ : str = self.tokenizer(self.input_text , return_tensors='''pt''' ) lowerCamelCase_ : int = model_abit_from_config.generate( input_ids=encoded_input['''input_ids'''].to(0 ) , max_new_tokens=1_0 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=A ) , self.EXPECTED_OUTPUTS ) def UpperCAmelCase__ (self ): with self.assertRaises(A ), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(A ) def UpperCAmelCase__ (self ): lowerCamelCase_ : List[Any] = BitsAndBytesConfig() with self.assertRaises(A ): lowerCamelCase_ : List[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=A , load_in_abit=A , device_map='''auto''' , bnb_abit_quant_type='''nf4''' , ) def UpperCAmelCase__ (self ): with self.assertRaises(A ): # Tries with `str` self.model_abit.to('''cpu''' ) with self.assertRaises(A ): # Tries with a `dtype`` self.model_abit.to(torch.floataa ) with self.assertRaises(A ): # Tries with a `device` self.model_abit.to(torch.device('''cuda:0''' ) ) with self.assertRaises(A ): # Tries with a `device` self.model_abit.float() with self.assertRaises(A ): # Tries with a `device` self.model_abit.half() # Test if we did not break anything lowerCamelCase_ : str = self.tokenizer(self.input_text , return_tensors='''pt''' ) lowerCamelCase_ : List[Any] = self.model_fpaa.to(torch.floataa ) lowerCamelCase_ : Tuple = self.model_fpaa.generate(input_ids=encoded_input['''input_ids'''].to(0 ) , max_new_tokens=1_0 ) # Check this does not throw an error lowerCamelCase_ : str = self.model_fpaa.to('''cpu''' ) # Check this does not throw an error lowerCamelCase_ : List[Any] = self.model_fpaa.half() # Check this does not throw an error lowerCamelCase_ : List[str] = self.model_fpaa.float() def UpperCAmelCase__ (self ): lowerCamelCase_ : str = AutoModelForSeqaSeqLM.from_pretrained('''t5-small''' , load_in_abit=A , device_map='''auto''' ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __lowercase ( unittest.TestCase ): @classmethod def UpperCAmelCase__ (cls ): lowerCamelCase_ : List[Any] = '''t5-small''' lowerCamelCase_ : Optional[Any] = '''google/flan-t5-small''' # flan-t5 uses dense-act instead of dense-relu-dense lowerCamelCase_ : List[str] = AutoTokenizer.from_pretrained(cls.model_name ) lowerCamelCase_ : Optional[Any] = '''Translate in German: Hello, my dog is cute''' def UpperCAmelCase__ (self ): gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ (self ): from transformers import TaForConditionalGeneration lowerCamelCase_ : Any = TaForConditionalGeneration._keep_in_fpaa_modules lowerCamelCase_ : List[Any] = None # test with `t5-small` lowerCamelCase_ : int = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=A , device_map='''auto''' ) lowerCamelCase_ : str = self.tokenizer(self.input_text , return_tensors='''pt''' ).to(0 ) lowerCamelCase_ : Optional[Any] = model.generate(**A ) # test with `flan-t5-small` lowerCamelCase_ : List[Any] = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=A , device_map='''auto''' ) lowerCamelCase_ : int = self.tokenizer(self.input_text , return_tensors='''pt''' ).to(0 ) lowerCamelCase_ : Optional[int] = model.generate(**A ) lowerCamelCase_ : Any = modules def UpperCAmelCase__ (self ): import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` lowerCamelCase_ : Tuple = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=A , device_map='''auto''' ) # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit ) ) lowerCamelCase_ : Optional[Any] = self.tokenizer(self.input_text , return_tensors='''pt''' ).to(0 ) lowerCamelCase_ : Dict = model.generate(**A ) # test with `flan-t5-small` lowerCamelCase_ : List[str] = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=A , device_map='''auto''' ) lowerCamelCase_ : Dict = self.tokenizer(self.input_text , return_tensors='''pt''' ).to(0 ) lowerCamelCase_ : Tuple = model.generate(**A ) class __lowercase ( _lowercase ): def UpperCAmelCase__ (self ): super().setUp() # model_name lowerCamelCase_ : Optional[int] = '''bigscience/bloom-560m''' lowerCamelCase_ : Optional[int] = '''t5-small''' # Different types of model lowerCamelCase_ : List[Any] = AutoModel.from_pretrained(self.model_name , load_in_abit=A , device_map='''auto''' ) # Sequence classification model lowerCamelCase_ : Dict = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=A , device_map='''auto''' ) # CausalLM model lowerCamelCase_ : int = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=A , device_map='''auto''' ) # Seq2seq model lowerCamelCase_ : int = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=A , device_map='''auto''' ) def UpperCAmelCase__ (self ): del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ (self ): from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit ) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter ) class __lowercase ( _lowercase ): def UpperCAmelCase__ (self ): super().setUp() def UpperCAmelCase__ (self ): del self.pipe gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ (self ): lowerCamelCase_ : int = pipeline( '''text-generation''' , model=self.model_name , model_kwargs={'''device_map''': '''auto''', '''load_in_4bit''': True, '''torch_dtype''': torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass lowerCamelCase_ : List[str] = self.pipe(self.input_text ) self.assertIn(pipeline_output[0]['''generated_text'''] , self.EXPECTED_OUTPUTS ) @require_torch_multi_gpu class __lowercase ( _lowercase ): def UpperCAmelCase__ (self ): super().setUp() def UpperCAmelCase__ (self ): lowerCamelCase_ : Optional[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=A , device_map='''balanced''' ) # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values() ) , {0, 1} ) # Check that inference pass works on the model lowerCamelCase_ : Any = self.tokenizer(self.input_text , return_tensors='''pt''' ) # Second real batch lowerCamelCase_ : Any = model_parallel.generate(input_ids=encoded_input['''input_ids'''].to(0 ) , max_new_tokens=1_0 ) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=A ) , self.EXPECTED_OUTPUTS ) class __lowercase ( _lowercase ): def UpperCAmelCase__ (self ): lowerCamelCase_ : str = '''facebook/opt-350m''' super().setUp() def UpperCAmelCase__ (self ): if version.parse(importlib.metadata.version('''bitsandbytes''' ) ) < version.parse('''0.37.0''' ): return # Step 1: freeze all parameters lowerCamelCase_ : Any = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=A ) self.assertEqual(set(model.hf_device_map.values() ) , {torch.cuda.current_device()} ) for param in model.parameters(): lowerCamelCase_ : List[str] = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability lowerCamelCase_ : Optional[int] = param.data.to(torch.floataa ) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(A ) ): lowerCamelCase_ : Dict = LoRALayer(module.q_proj , rank=1_6 ) lowerCamelCase_ : str = LoRALayer(module.k_proj , rank=1_6 ) lowerCamelCase_ : int = LoRALayer(module.v_proj , rank=1_6 ) # Step 3: dummy batch lowerCamelCase_ : Union[str, Any] = self.tokenizer('''Test batch ''' , return_tensors='''pt''' ).to(0 ) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): lowerCamelCase_ : Optional[int] = model.forward(**A ) out.logits.norm().backward() for module in model.modules(): if isinstance(A , A ): self.assertTrue(module.adapter[1].weight.grad is not None ) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0 ) elif isinstance(A , nn.Embedding ): self.assertTrue(module.weight.grad is None ) class __lowercase ( _lowercase ): lowerCamelCase : Optional[Any] = "gpt2-xl" lowerCamelCase : int = 3.3_1_9_1_8_5_4_8_5_4_1_5_2_1_8_7
318
0
from typing import TYPE_CHECKING from ...utils import _LazyModule A : Optional[int] = {'tokenization_byt5': ['ByT5Tokenizer']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys A : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
146
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class A : '''simple docstring''' def __init__(self : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : Any=13 , _UpperCAmelCase : int=64 , _UpperCAmelCase : Any=2 , _UpperCAmelCase : str=3 , _UpperCAmelCase : str=True , _UpperCAmelCase : List[Any]=True , _UpperCAmelCase : Optional[int]=32 , _UpperCAmelCase : Tuple=5 , _UpperCAmelCase : Tuple=4 , _UpperCAmelCase : Union[str, Any]=37 , _UpperCAmelCase : Tuple="gelu" , _UpperCAmelCase : Optional[int]=0.1 , _UpperCAmelCase : int=0.1 , _UpperCAmelCase : Any=10 , _UpperCAmelCase : str=0.02 , _UpperCAmelCase : Any=[1, 16, 4, 4] , _UpperCAmelCase : str=None , ) -> int: """simple docstring""" lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = is_training lowercase__ = use_labels lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = scope lowercase__ = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size lowercase__ = (self.image_size // 32) ** 2 lowercase__ = num_patches + 1 def lowerCamelCase__ (self : Any ) -> Optional[int]: """simple docstring""" lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = self.get_config() return config, pixel_values, labels def lowerCamelCase__ (self : int ) -> Tuple: """simple docstring""" lowercase__ = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, """hidden_sizes""": [4, 8, 16, 32], """num_groups""": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_UpperCAmelCase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=_UpperCAmelCase , ) def lowerCamelCase__ (self : Union[str, Any] , _UpperCAmelCase : int , _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> Optional[Any]: """simple docstring""" lowercase__ = ViTHybridModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() lowercase__ = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ (self : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : str , _UpperCAmelCase : int ) -> List[Any]: """simple docstring""" lowercase__ = self.type_sequence_label_size lowercase__ = ViTHybridForImageClassification(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() lowercase__ = model(_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase__ (self : int ) -> Union[str, Any]: """simple docstring""" lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class A ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' A__ = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () A__ = ( {'''feature-extraction''': ViTHybridModel, '''image-classification''': ViTHybridForImageClassification} if is_torch_available() else {} ) A__ = False A__ = False A__ = False def lowerCamelCase__ (self : Any ) -> Any: """simple docstring""" lowercase__ = ViTHybridModelTester(self ) lowercase__ = ConfigTester(self , config_class=_UpperCAmelCase , has_text_modality=_UpperCAmelCase , hidden_size=37 ) def lowerCamelCase__ (self : Dict ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def lowerCamelCase__ (self : List[str] ) -> Any: """simple docstring""" pass def lowerCamelCase__ (self : Dict ) -> Union[str, Any]: """simple docstring""" lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_UpperCAmelCase , nn.Linear ) ) def lowerCamelCase__ (self : Dict ) -> Optional[Any]: """simple docstring""" lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_UpperCAmelCase ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _UpperCAmelCase ) def lowerCamelCase__ (self : Optional[Any] ) -> Dict: """simple docstring""" lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def lowerCamelCase__ (self : Tuple ) -> Optional[Any]: """simple docstring""" lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_UpperCAmelCase ) def lowerCamelCase__ (self : Dict ) -> Any: """simple docstring""" lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = _config_zero_init(_UpperCAmelCase ) for model_class in self.all_model_classes: lowercase__ = model_class(config=_UpperCAmelCase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": lowercase__ = [f'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @slow def lowerCamelCase__ (self : Dict ) -> List[Any]: """simple docstring""" for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = ViTHybridModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def UpperCamelCase ( ) -> Optional[int]: """simple docstring""" lowercase__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class A ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCamelCase__ (self : Optional[int] ) -> Union[str, Any]: """simple docstring""" return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCamelCase__ (self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowercase__ = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( _UpperCAmelCase ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_UpperCAmelCase , return_tensors="""pt""" ).to(_UpperCAmelCase ) # forward pass with torch.no_grad(): lowercase__ = model(**_UpperCAmelCase ) # verify the logits lowercase__ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _UpperCAmelCase ) lowercase__ = torch.tensor([-1.9_090, -0.4_993, -0.2_389] ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _UpperCAmelCase , atol=1E-4 ) ) @slow @require_accelerate def lowerCamelCase__ (self : Any ) -> Optional[int]: """simple docstring""" lowercase__ = ViTHybridImageProcessor.from_pretrained("""google/vit-hybrid-base-bit-384""" ) lowercase__ = ViTHybridForImageClassification.from_pretrained("""google/vit-hybrid-base-bit-384""" , device_map="""auto""" ) lowercase__ = prepare_img() lowercase__ = image_processor(images=_UpperCAmelCase , return_tensors="""pt""" ) lowercase__ = model(**_UpperCAmelCase ) lowercase__ = outputs.logits # model predicts one of the 1000 ImageNet classes lowercase__ = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , """tabby, tabby cat""" )
146
1
"""simple docstring""" import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __A : Union[str, Any] = "\\n\n" __A : Any = "\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n" __A : List[str] = "\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to 'cuda' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"]\n >>> results = perplexity.compute(model_id='gpt2',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 78.22\n >>> print(round(results[\"perplexities\"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = datasets.load_dataset(\"wikitext\",\n ... \"wikitext-2-raw-v1\",\n ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!='']\n >>> results = perplexity.compute(model_id='gpt2',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 60.35\n >>> print(round(results[\"perplexities\"][0], 2))\n 81.12\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _a ( datasets.Metric): """simple docstring""" def lowercase__ ( self : List[Any] )->Union[str, Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''input_texts''': datasets.Value('''string''' ), } ) , reference_urls=['''https://huggingface.co/docs/transformers/perplexity'''] , ) def lowercase__ ( self : Dict , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Dict , __UpperCamelCase : int = 1_6 , __UpperCamelCase : bool = True , __UpperCamelCase : List[Any]=None )->Any: if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": _UpperCAmelCase = '''cuda''' else: _UpperCAmelCase = '''cuda''' if torch.cuda.is_available() else '''cpu''' _UpperCAmelCase = AutoModelForCausalLM.from_pretrained(__UpperCamelCase ) _UpperCAmelCase = model.to(__UpperCamelCase ) _UpperCAmelCase = AutoTokenizer.from_pretrained(__UpperCamelCase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: _UpperCAmelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(__UpperCamelCase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'''pad_token''': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" _UpperCAmelCase = model.config.max_length - 1 else: _UpperCAmelCase = model.config.max_length _UpperCAmelCase = tokenizer( __UpperCamelCase , add_special_tokens=__UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase , max_length=__UpperCamelCase , return_tensors='''pt''' , return_attention_mask=__UpperCamelCase , ).to(__UpperCamelCase ) _UpperCAmelCase = encodings['''input_ids'''] _UpperCAmelCase = encodings['''attention_mask'''] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." _UpperCAmelCase = [] _UpperCAmelCase = CrossEntropyLoss(reduction='''none''' ) for start_index in logging.tqdm(range(0 , len(__UpperCamelCase ) , __UpperCamelCase ) ): _UpperCAmelCase = min(start_index + batch_size , len(__UpperCamelCase ) ) _UpperCAmelCase = encoded_texts[start_index:end_index] _UpperCAmelCase = attn_masks[start_index:end_index] if add_start_token: _UpperCAmelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(__UpperCamelCase ) _UpperCAmelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) _UpperCAmelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(__UpperCamelCase ), attn_mask] , dim=1 ) _UpperCAmelCase = encoded_batch with torch.no_grad(): _UpperCAmelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase ).logits _UpperCAmelCase = out_logits[..., :-1, :].contiguous() _UpperCAmelCase = labels[..., 1:].contiguous() _UpperCAmelCase = attn_mask[..., 1:].contiguous() _UpperCAmelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , __UpperCamelCase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(__UpperCamelCase )}
260
"""simple docstring""" from pathlib import Path import numpy as np from PIL import Image def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2989 * r + 0.5870 * g + 0.1140 * b def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' return (gray > 127) & (gray <= 255) def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' _UpperCAmelCase = np.zeros_like(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image _UpperCAmelCase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): _UpperCAmelCase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() _UpperCAmelCase = int(summation > 0 ) return output if __name__ == "__main__": # read original image __A : str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" __A : str = np.array(Image.open(lena_path)) # kernel to be applied __A : List[Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) __A : Optional[Any] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image __A : Optional[Any] = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
260
1
import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) SCREAMING_SNAKE_CASE : Dict = [ "cross_validation.py", "gradient_accumulation.py", "local_sgd.py", "multi_process_metrics.py", "memory.py", "automatic_gradient_accumulation.py", "fsdp_with_peak_mem_tracking.py", "deepspeed_with_config_support.py", "megatron_lm_gpt_pretraining.py", ] class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None ): lowercase_ :Union[str, Any] = None lowercase_ :List[str] = os.path.abspath(os.path.join('''examples''' , '''by_feature''' ) ) lowercase_ :int = os.path.abspath('''examples''' ) for item in os.listdir(UpperCamelCase_ ): if item not in EXCLUDE_EXAMPLES: lowercase_ :Optional[Any] = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) if os.path.isfile(UpperCamelCase_ ) and ".py" in item_path: with self.subTest( tested_script=UpperCamelCase_ , feature_script=UpperCamelCase_ , tested_section='''main()''' if parser_only else '''training_function()''' , ): lowercase_ :Optional[int] = compare_against_test( os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) lowercase_ :Optional[Any] = '''\n'''.join(UpperCamelCase_ ) if special_strings is not None: for string in special_strings: lowercase_ :str = diff.replace(UpperCamelCase_ , '''''' ) self.assertEqual(UpperCamelCase_ , '''''' ) def UpperCamelCase ( self ): self.one_complete_example('''complete_nlp_example.py''' , UpperCamelCase_ ) self.one_complete_example('''complete_nlp_example.py''' , UpperCamelCase_ ) def UpperCamelCase ( self ): lowercase_ :int = os.path.abspath(os.path.join('''examples''' , '''cv_example.py''' ) ) lowercase_ :Optional[int] = [ ''' ''' * 16 + '''{\n\n''', ''' ''' * 20 + '''"accuracy": eval_metric["accuracy"],\n\n''', ''' ''' * 20 + '''"f1": eval_metric["f1"],\n\n''', ''' ''' * 20 + '''"train_loss": total_loss.item() / len(train_dataloader),\n\n''', ''' ''' * 20 + '''"epoch": epoch,\n\n''', ''' ''' * 16 + '''},\n\n''', ''' ''' * 16 + '''step=epoch,\n''', ''' ''' * 12, ''' ''' * 8 + '''for step, batch in enumerate(active_dataloader):\n''', ] self.one_complete_example('''complete_cv_example.py''' , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) self.one_complete_example('''complete_cv_example.py''' , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) @mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """1"""} ) class UpperCamelCase ( lowercase__ ): '''simple docstring''' lowercase : str =False @classmethod def UpperCamelCase ( cls ): super().setUpClass() lowercase_ :List[Any] = tempfile.mkdtemp() lowercase_ :Dict = os.path.join(cls._tmpdir , '''default_config.yml''' ) write_basic_config(save_location=cls.configPath ) lowercase_ :Dict = ['''accelerate''', '''launch''', '''--config_file''', cls.configPath] @classmethod def UpperCamelCase ( cls ): super().tearDownClass() shutil.rmtree(cls._tmpdir ) def UpperCamelCase ( self ): lowercase_ :Tuple = f"\n examples/by_feature/checkpointing.py\n --checkpointing_steps epoch\n --output_dir {self.tmpdir}\n ".split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , '''epoch_0''' ) ) ) def UpperCamelCase ( self ): lowercase_ :Optional[int] = f"\n examples/by_feature/checkpointing.py\n --checkpointing_steps 1\n --output_dir {self.tmpdir}\n ".split() lowercase_ :str = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , '''step_2''' ) ) ) def UpperCamelCase ( self ): lowercase_ :str = f"\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , 'epoch_0' )}\n ".split() lowercase_ :Optional[Any] = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase_ ) self.assertNotIn('''epoch 0:''' , UpperCamelCase_ ) self.assertIn('''epoch 1:''' , UpperCamelCase_ ) def UpperCamelCase ( self ): lowercase_ :str = f"\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , 'step_2' )}\n ".split() lowercase_ :Union[str, Any] = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase_ ) if torch.cuda.is_available(): lowercase_ :Any = torch.cuda.device_count() else: lowercase_ :int = 1 if num_processes > 1: self.assertNotIn('''epoch 0:''' , UpperCamelCase_ ) self.assertIn('''epoch 1:''' , UpperCamelCase_ ) else: self.assertIn('''epoch 0:''' , UpperCamelCase_ ) self.assertIn('''epoch 1:''' , UpperCamelCase_ ) @slow def UpperCamelCase ( self ): lowercase_ :Dict = ''' examples/by_feature/cross_validation.py --num_folds 2 '''.split() with mock.patch.dict(os.environ , {'''TESTING_MOCKED_DATALOADERS''': '''0'''} ): lowercase_ :Union[str, Any] = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase_ ) lowercase_ :Optional[Any] = re.findall('''({.+})''' , UpperCamelCase_ ) lowercase_ :List[Any] = [r for r in results if '''accuracy''' in r][-1] lowercase_ :str = ast.literal_eval(UpperCamelCase_ ) self.assertGreaterEqual(results['''accuracy'''] , 0.75 ) def UpperCamelCase ( self ): lowercase_ :List[str] = ['''examples/by_feature/multi_process_metrics.py'''] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: lowercase_ :str = f"\n examples/by_feature/tracking.py\n --with_tracking\n --project_dir {tmpdir}\n ".split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase_ , '''tracking''' ) ) ) def UpperCamelCase ( self ): lowercase_ :int = ['''examples/by_feature/gradient_accumulation.py'''] run_command(self._launch_args + testargs ) def UpperCamelCase ( self ): lowercase_ :str = ['''examples/by_feature/local_sgd.py'''] run_command(self._launch_args + testargs )
252
import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase ( self ): lowercase_ :int = logging.get_logger() # the current default level is logging.WARNING lowercase_ :List[str] = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(UpperCamelCase_ ) def UpperCamelCase ( self ): lowercase_ :Tuple = logging.get_verbosity() lowercase_ :str = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Tuple = '''Testing 1, 2, 3''' # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , '''''' ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) # restore to the original level logging.set_verbosity(UpperCamelCase_ ) @mockenv(TRANSFORMERS_VERBOSITY='''error''' ) def UpperCamelCase ( self ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() # this action activates the env var lowercase_ :Any = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Optional[Any] = os.getenv('''TRANSFORMERS_VERBOSITY''' , UpperCamelCase_ ) lowercase_ :Any = logging.log_levels[env_level_str] lowercase_ :Optional[int] = logging.get_verbosity() self.assertEqual( UpperCamelCase_ , UpperCamelCase_ , f"TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}" , ) # restore to the original level lowercase_ :str = '''''' transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY='''super-error''' ) def UpperCamelCase ( self ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() lowercase_ :Any = logging.logging.getLogger() with CaptureLogger(UpperCamelCase_ ) as cl: # this action activates the env var logging.get_logger('''transformers.models.bart.tokenization_bart''' ) self.assertIn('''Unknown option TRANSFORMERS_VERBOSITY=super-error''' , cl.out ) # no need to restore as nothing was changed def UpperCamelCase ( self ): # testing `logger.warning_advice()` transformers.utils.logging._reset_library_root_logger() lowercase_ :Optional[int] = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Any = '''Testing 1, 2, 3''' with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''1''' ): # nothing should be logged as env var disables this method with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning_advice(UpperCamelCase_ ) self.assertEqual(cl.out , '''''' ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''''' ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning_advice(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
252
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = '''▁''' UpperCAmelCase__ = {'''vocab_file''': '''spiece.model'''} UpperCAmelCase__ = { '''vocab_file''': { '''google/reformer-crime-and-punishment''': ( '''https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model''' ) } } UpperCAmelCase__ = { '''google/reformer-crime-and-punishment''': 52_4288, } class lowerCamelCase__ ( lowerCAmelCase): SCREAMING_SNAKE_CASE__ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ = ['''input_ids''', '''attention_mask'''] def __init__(self , UpperCAmelCase , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=[] , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: _lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase , ) _lowercase =vocab_file _lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase ) @property def __A (self ) -> Optional[int]: return self.sp_model.get_piece_size() def __A (self ) -> Dict[str, int]: _lowercase ={self.convert_ids_to_tokens(UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__(self ) -> Union[str, Any]: _lowercase =self.__dict__.copy() _lowercase =None return state def __setstate__(self , UpperCAmelCase ) -> Optional[Any]: _lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _lowercase ={} _lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __A (self , UpperCAmelCase ) -> List[str]: return self.sp_model.encode(UpperCAmelCase , out_type=UpperCAmelCase ) def __A (self , UpperCAmelCase ) -> str: return self.sp_model.piece_to_id(UpperCAmelCase ) def __A (self , UpperCAmelCase ) -> Tuple: if index < self.sp_model.get_piece_size(): _lowercase =self.sp_model.IdToPiece(UpperCAmelCase ) return token def __A (self , UpperCAmelCase ) -> str: _lowercase =[] _lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCAmelCase ) + token _lowercase =[] else: current_sub_tokens.append(UpperCAmelCase ) out_string += self.sp_model.decode(UpperCAmelCase ) return out_string.strip() def __A (self , UpperCAmelCase , UpperCAmelCase = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _lowercase =os.path.join( UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase , '''wb''' ) as fi: _lowercase =self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase ) return (out_vocab_file,)
5
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : Tuple = """▁""" UpperCAmelCase_ : Optional[Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} UpperCAmelCase_ : str = { """vocab_file""": { """facebook/xglm-564M""": """https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model""", } } UpperCAmelCase_ : str = { """facebook/xglm-564M""": 2048, } class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ["input_ids", "attention_mask"] def __init__( self : List[Any] , lowercase_ : str , lowercase_ : Tuple="<s>" , lowercase_ : Any="</s>" , lowercase_ : Optional[int]="</s>" , lowercase_ : List[Any]="<s>" , lowercase_ : Union[str, Any]="<unk>" , lowercase_ : Union[str, Any]="<pad>" , lowercase_ : Optional[Dict[str, Any]] = None , **lowercase_ : Tuple , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Any = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer SCREAMING_SNAKE_CASE_ : List[str] = 7 SCREAMING_SNAKE_CASE_ : Tuple = [F'<madeupword{i}>' for i in range(self.num_madeup_words)] SCREAMING_SNAKE_CASE_ : List[Any] = kwargs.get('''additional_special_tokens''' , []) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , pad_token=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , **lowercase_ , ) SCREAMING_SNAKE_CASE_ : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowercase_)) SCREAMING_SNAKE_CASE_ : Union[str, Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab SCREAMING_SNAKE_CASE_ : Union[str, Any] = 1 # Mimic fairseq token-to-id alignment for the first 4 token SCREAMING_SNAKE_CASE_ : Optional[Any] = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} SCREAMING_SNAKE_CASE_ : List[Any] = len(self.sp_model) SCREAMING_SNAKE_CASE_ : Optional[Any] = {F'<madeupword{i}>': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words)} self.fairseq_tokens_to_ids.update(lowercase_) SCREAMING_SNAKE_CASE_ : List[Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Dict): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Any = self.__dict__.copy() SCREAMING_SNAKE_CASE_ : str = None SCREAMING_SNAKE_CASE_ : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , lowercase_ : str): '''simple docstring''' SCREAMING_SNAKE_CASE_ : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): SCREAMING_SNAKE_CASE_ : Union[str, Any] = {} SCREAMING_SNAKE_CASE_ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None): '''simple docstring''' if token_ids_a is None: return [self.sep_token_id] + token_ids_a SCREAMING_SNAKE_CASE_ : Dict = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) if token_ids_a is None: return [1] + ([0] * len(lowercase_)) return [1] + ([0] * len(lowercase_)) + [1, 1] + ([0] * len(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Union[str, Any] = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a) * [0] @property def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): '''simple docstring''' return len(self.sp_model) + self.fairseq_offset + self.num_madeup_words def _SCREAMING_SNAKE_CASE ( self : Any): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def _SCREAMING_SNAKE_CASE ( self : Dict , lowercase_ : str): '''simple docstring''' return self.sp_model.encode(lowercase_ , out_type=lowercase_) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowercase_ : Union[str, Any]): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] SCREAMING_SNAKE_CASE_ : Optional[Any] = self.sp_model.PieceToId(lowercase_) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : Optional[Any]): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : Tuple): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = ''''''.join(lowercase_).replace(lowercase_ , ''' ''').strip() return out_string def _SCREAMING_SNAKE_CASE ( self : str , lowercase_ : str , lowercase_ : Optional[str] = None): '''simple docstring''' if not os.path.isdir(lowercase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return SCREAMING_SNAKE_CASE_ : List[Any] = os.path.join( lowercase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , '''wb''') as fi: SCREAMING_SNAKE_CASE_ : int = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
91
0
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class lowerCAmelCase__ ( unittest.TestCase ): def __a ( self : Optional[Any] , snake_case__ : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = 3 UpperCAmelCase__ : List[Any] = 2_5_0 UpperCAmelCase__ : List[str] = ids_tensor((batch_size, length) , snake_case__ ) UpperCAmelCase__ : Any = torch.ones((batch_size, length) , device=snake_case__ , dtype=torch.float ) / length return input_ids, scores def __a ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self._get_tensors(5 ) UpperCAmelCase__ : List[Any] = StoppingCriteriaList( [ MaxLengthCriteria(max_length=1_0 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : int = self._get_tensors(9 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : int = self._get_tensors(1_0 ) self.assertTrue(criteria(snake_case__ , snake_case__ ) ) def __a ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaxLengthCriteria(max_length=1_0 ) UpperCAmelCase__ : Optional[int] = self._get_tensors(5 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : Union[str, Any] = self._get_tensors(9 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : Optional[int] = self._get_tensors(1_0 ) self.assertTrue(criteria(snake_case__ , snake_case__ ) ) def __a ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : str = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) UpperCAmelCase__ : Optional[int] = self._get_tensors(5 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : List[Any] = self._get_tensors(9 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : Any = self._get_tensors(1_0 ) self.assertTrue(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : List[str] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 1_0 ) def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : int = self._get_tensors(5 ) UpperCAmelCase__ : Union[str, Any] = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(snake_case__ , snake_case__ ) ) UpperCAmelCase__ : Union[str, Any] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(snake_case__ , snake_case__ ) ) def __a ( self : str ): '''simple docstring''' validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_0 ) with self.assertWarns(snake_case__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_1 ) UpperCAmelCase__ : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 1_1 ) self.assertEqual(len(snake_case__ ) , 1 )
354
"""simple docstring""" import numpy as np import datasets _lowerCAmelCase : Optional[int] = """ Compute the Mahalanobis Distance Mahalonobis distance is the distance between a point and a distribution. And not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance. It was introduced by Prof. P. C. Mahalanobis in 1936 and has been used in various statistical applications ever since [source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/] """ _lowerCAmelCase : Tuple = """\ @article{de2000mahalanobis, title={The mahalanobis distance}, author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L}, journal={Chemometrics and intelligent laboratory systems}, volume={50}, number={1}, pages={1--18}, year={2000}, publisher={Elsevier} } """ _lowerCAmelCase : Optional[int] = """ Args: X: List of datapoints to be compared with the `reference_distribution`. reference_distribution: List of datapoints from the reference distribution we want to compare to. Returns: mahalanobis: The Mahalonobis distance for each datapoint in `X`. Examples: >>> mahalanobis_metric = datasets.load_metric(\"mahalanobis\") >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]]) >>> print(results) {'mahalanobis': array([0.5])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def __a ( self : Any ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence" ) , id="X" ), } ) , ) def __a ( self : Union[str, Any] , snake_case__ : List[str] , snake_case__ : Any ): '''simple docstring''' # convert to numpy arrays UpperCAmelCase__ : Union[str, Any] = np.array(snake_case__ ) UpperCAmelCase__ : Union[str, Any] = np.array(snake_case__ ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError("Expected `X` to be a 2D vector" ) if len(reference_distribution.shape ) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector" ) if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension" ) # Get mahalanobis distance for each prediction UpperCAmelCase__ : Optional[Any] = X - np.mean(snake_case__ ) UpperCAmelCase__ : Tuple = np.cov(reference_distribution.T ) try: UpperCAmelCase__ : str = np.linalg.inv(snake_case__ ) except np.linalg.LinAlgError: UpperCAmelCase__ : Optional[Any] = np.linalg.pinv(snake_case__ ) UpperCAmelCase__ : List[Any] = np.dot(snake_case__ , snake_case__ ) UpperCAmelCase__ : Tuple = np.dot(snake_case__ , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
298
0
'''simple docstring''' from random import randint, random def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = 5 , )-> list: """simple docstring""" __A = [[-1] * number_of_cells] # Create a highway without any car __A = 0 __A = max(UpperCAmelCase , 0 ) while i < number_of_cells: __A = ( randint(0 , UpperCAmelCase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def snake_case ( UpperCAmelCase , UpperCAmelCase )-> int: """simple docstring""" __A = 0 __A = highway_now[car_index + 1 :] for cell in range(len(UpperCAmelCase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(UpperCAmelCase , -1 ) def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )-> list: """simple docstring""" __A = len(UpperCAmelCase ) # Beforce calculations, the highway is empty __A = [-1] * number_of_cells for car_index in range(UpperCAmelCase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed __A = min(highway_now[car_index] + 1 , UpperCAmelCase ) # Number of empty cell before the next car __A = get_distance(UpperCAmelCase , UpperCAmelCase ) - 1 # We can't have the car causing an accident __A = min(next_highway[car_index] , UpperCAmelCase ) if random() < probability: # Randomly, a driver will slow down __A = max(next_highway[car_index] - 1 , 0 ) return next_highway def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )-> list: """simple docstring""" __A = len(highway[0] ) for i in range(UpperCAmelCase ): __A = update(highway[i] , UpperCAmelCase , UpperCAmelCase ) __A = [-1] * number_of_cells for car_index in range(UpperCAmelCase ): __A = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) __A = (car_index + speed) % number_of_cells # Commit the change of position __A = speed highway.append(UpperCAmelCase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
161
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a__ : List[Any] = { "configuration_vision_text_dual_encoder": ["VisionTextDualEncoderConfig"], "processing_vision_text_dual_encoder": ["VisionTextDualEncoderProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Tuple = ["VisionTextDualEncoderModel"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Union[str, Any] = ["FlaxVisionTextDualEncoderModel"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] = ["TFVisionTextDualEncoderModel"] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys a__ : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure)
161
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : int = { """xlm-mlm-en-2048""": """https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json""", """xlm-mlm-ende-1024""": """https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json""", """xlm-mlm-enfr-1024""": """https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json""", """xlm-mlm-enro-1024""": """https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json""", """xlm-mlm-tlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json""", """xlm-mlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json""", """xlm-clm-enfr-1024""": """https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json""", """xlm-clm-ende-1024""": """https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json""", """xlm-mlm-17-1280""": """https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json""", """xlm-mlm-100-1280""": """https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json""", } class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = "xlm" __UpperCamelCase = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self : str , lowercase_ : int=30145 , lowercase_ : List[str]=2048 , lowercase_ : Dict=12 , lowercase_ : int=16 , lowercase_ : List[Any]=0.1 , lowercase_ : Any=0.1 , lowercase_ : List[str]=True , lowercase_ : Optional[int]=False , lowercase_ : Any=False , lowercase_ : List[str]=False , lowercase_ : List[Any]=1 , lowercase_ : Optional[Any]=True , lowercase_ : int=512 , lowercase_ : Dict=2048**-0.5 , lowercase_ : Any=1e-12 , lowercase_ : str=0.02 , lowercase_ : str=0 , lowercase_ : Optional[Any]=1 , lowercase_ : Tuple=2 , lowercase_ : Dict=3 , lowercase_ : List[Any]=5 , lowercase_ : List[str]=True , lowercase_ : Any="first" , lowercase_ : Union[str, Any]=True , lowercase_ : str=None , lowercase_ : Tuple=True , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[Any]=5 , lowercase_ : Optional[int]=5 , lowercase_ : int=0 , lowercase_ : Optional[Any]=0 , lowercase_ : Dict=2 , lowercase_ : Optional[Any]=0 , **lowercase_ : Tuple , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Union[str, Any] = vocab_size SCREAMING_SNAKE_CASE_ : List[Any] = emb_dim SCREAMING_SNAKE_CASE_ : Any = n_layers SCREAMING_SNAKE_CASE_ : Optional[int] = n_heads SCREAMING_SNAKE_CASE_ : Any = dropout SCREAMING_SNAKE_CASE_ : Any = attention_dropout SCREAMING_SNAKE_CASE_ : List[str] = gelu_activation SCREAMING_SNAKE_CASE_ : Optional[Any] = sinusoidal_embeddings SCREAMING_SNAKE_CASE_ : str = causal SCREAMING_SNAKE_CASE_ : str = asm SCREAMING_SNAKE_CASE_ : Tuple = n_langs SCREAMING_SNAKE_CASE_ : Any = use_lang_emb SCREAMING_SNAKE_CASE_ : Optional[Any] = layer_norm_eps SCREAMING_SNAKE_CASE_ : str = bos_index SCREAMING_SNAKE_CASE_ : List[str] = eos_index SCREAMING_SNAKE_CASE_ : List[Any] = pad_index SCREAMING_SNAKE_CASE_ : int = unk_index SCREAMING_SNAKE_CASE_ : int = mask_index SCREAMING_SNAKE_CASE_ : Dict = is_encoder SCREAMING_SNAKE_CASE_ : str = max_position_embeddings SCREAMING_SNAKE_CASE_ : List[Any] = embed_init_std SCREAMING_SNAKE_CASE_ : Optional[int] = init_std SCREAMING_SNAKE_CASE_ : Dict = summary_type SCREAMING_SNAKE_CASE_ : Tuple = summary_use_proj SCREAMING_SNAKE_CASE_ : int = summary_activation SCREAMING_SNAKE_CASE_ : Optional[Any] = summary_proj_to_labels SCREAMING_SNAKE_CASE_ : Any = summary_first_dropout SCREAMING_SNAKE_CASE_ : str = start_n_top SCREAMING_SNAKE_CASE_ : Optional[Any] = end_n_top SCREAMING_SNAKE_CASE_ : Tuple = mask_token_id SCREAMING_SNAKE_CASE_ : Dict = lang_id if "n_words" in kwargs: SCREAMING_SNAKE_CASE_ : Dict = kwargs['''n_words'''] super().__init__(pad_token_id=lowercase_ , bos_token_id=lowercase_ , **lowercase_) class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self : Tuple): '''simple docstring''' if self.task == "multiple-choice": SCREAMING_SNAKE_CASE_ : int = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: SCREAMING_SNAKE_CASE_ : List[str] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ])
357
"""simple docstring""" from ...processing_utils import ProcessorMixin class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = "SpeechT5FeatureExtractor" __UpperCamelCase = "SpeechT5Tokenizer" def __init__( self : Any , lowercase_ : Dict , lowercase_ : Optional[Any]): '''simple docstring''' super().__init__(lowercase_ , lowercase_) def __call__( self : List[Any] , *lowercase_ : List[Any] , **lowercase_ : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = kwargs.pop('''audio''' , lowercase_) SCREAMING_SNAKE_CASE_ : Optional[int] = kwargs.pop('''text''' , lowercase_) SCREAMING_SNAKE_CASE_ : Any = kwargs.pop('''text_target''' , lowercase_) SCREAMING_SNAKE_CASE_ : Dict = kwargs.pop('''audio_target''' , lowercase_) SCREAMING_SNAKE_CASE_ : List[Any] = kwargs.pop('''sampling_rate''' , lowercase_) if audio is not None and text is not None: raise ValueError( '''Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?''') if audio_target is not None and text_target is not None: raise ValueError( '''Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?''') if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( '''You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.''') if audio is not None: SCREAMING_SNAKE_CASE_ : Any = self.feature_extractor(lowercase_ , *lowercase_ , sampling_rate=lowercase_ , **lowercase_) elif text is not None: SCREAMING_SNAKE_CASE_ : Dict = self.tokenizer(lowercase_ , **lowercase_) else: SCREAMING_SNAKE_CASE_ : Any = None if audio_target is not None: SCREAMING_SNAKE_CASE_ : List[Any] = self.feature_extractor(audio_target=lowercase_ , *lowercase_ , sampling_rate=lowercase_ , **lowercase_) SCREAMING_SNAKE_CASE_ : Optional[int] = targets['''input_values'''] elif text_target is not None: SCREAMING_SNAKE_CASE_ : int = self.tokenizer(lowercase_ , **lowercase_) SCREAMING_SNAKE_CASE_ : List[Any] = targets['''input_ids'''] else: SCREAMING_SNAKE_CASE_ : int = None if inputs is None: return targets if targets is not None: SCREAMING_SNAKE_CASE_ : Union[str, Any] = labels SCREAMING_SNAKE_CASE_ : Optional[Any] = targets.get('''attention_mask''') if decoder_attention_mask is not None: SCREAMING_SNAKE_CASE_ : Any = decoder_attention_mask return inputs def _SCREAMING_SNAKE_CASE ( self : Tuple , *lowercase_ : Tuple , **lowercase_ : List[str]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = kwargs.pop('''input_values''' , lowercase_) SCREAMING_SNAKE_CASE_ : int = kwargs.pop('''input_ids''' , lowercase_) SCREAMING_SNAKE_CASE_ : Dict = kwargs.pop('''labels''' , lowercase_) if input_values is not None and input_ids is not None: raise ValueError('''Cannot process both `input_values` and `input_ids` inputs.''') if input_values is None and input_ids is None and labels is None: raise ValueError( '''You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.''') if input_values is not None: SCREAMING_SNAKE_CASE_ : Any = self.feature_extractor.pad(lowercase_ , *lowercase_ , **lowercase_) elif input_ids is not None: SCREAMING_SNAKE_CASE_ : Tuple = self.tokenizer.pad(lowercase_ , **lowercase_) else: SCREAMING_SNAKE_CASE_ : List[Any] = None if labels is not None: if "input_ids" in labels or (isinstance(lowercase_ , lowercase_) and "input_ids" in labels[0]): SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.tokenizer.pad(lowercase_ , **lowercase_) SCREAMING_SNAKE_CASE_ : Dict = targets['''input_ids'''] else: SCREAMING_SNAKE_CASE_ : Dict = self.feature_extractor.feature_size SCREAMING_SNAKE_CASE_ : Optional[int] = self.feature_extractor.num_mel_bins SCREAMING_SNAKE_CASE_ : str = self.feature_extractor.pad(lowercase_ , *lowercase_ , **lowercase_) SCREAMING_SNAKE_CASE_ : str = feature_size_hack SCREAMING_SNAKE_CASE_ : Dict = targets['''input_values'''] else: SCREAMING_SNAKE_CASE_ : List[Any] = None if inputs is None: return targets if targets is not None: SCREAMING_SNAKE_CASE_ : Dict = labels SCREAMING_SNAKE_CASE_ : List[str] = targets.get('''attention_mask''') if decoder_attention_mask is not None: SCREAMING_SNAKE_CASE_ : Optional[Any] = decoder_attention_mask return inputs def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , *lowercase_ : Optional[int] , **lowercase_ : Tuple): '''simple docstring''' return self.tokenizer.batch_decode(*lowercase_ , **lowercase_) def _SCREAMING_SNAKE_CASE ( self : List[Any] , *lowercase_ : Dict , **lowercase_ : List[Any]): '''simple docstring''' return self.tokenizer.decode(*lowercase_ , **lowercase_)
318
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __SCREAMING_SNAKE_CASE ( lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaImgaImgPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """image"""] snake_case_ = [ """image_embeds""", """negative_image_embeds""", """image""", ] snake_case_ = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __magic_name__ ( self : str ) -> List[str]: return 32 @property def __magic_name__ ( self : Union[str, Any] ) -> List[str]: return 32 @property def __magic_name__ ( self : Optional[int] ) -> str: return self.time_input_dim @property def __magic_name__ ( self : int ) -> Optional[int]: return self.time_input_dim * 4 @property def __magic_name__ ( self : Union[str, Any] ) -> Dict: return 1_00 @property def __magic_name__ ( self : Any ) -> Tuple: torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Tuple ={ '''in_channels''': 4, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } SCREAMING_SNAKE_CASE__ : Optional[int] =UNetaDConditionModel(**__lowercase ) return model @property def __magic_name__ ( self : Dict ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __magic_name__ ( self : List[Any] ) -> str: torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : List[str] =VQModel(**self.dummy_movq_kwargs ) return model def __magic_name__ ( self : List[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Optional[Any] =self.dummy_unet SCREAMING_SNAKE_CASE__ : str =self.dummy_movq SCREAMING_SNAKE_CASE__ : List[str] ={ '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } SCREAMING_SNAKE_CASE__ : Tuple =DDIMScheduler(**__lowercase ) SCREAMING_SNAKE_CASE__ : int ={ '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def __magic_name__ ( self : Optional[Any] , __lowercase : Optional[int] , __lowercase : Dict=0 ) -> Any: SCREAMING_SNAKE_CASE__ : List[Any] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image SCREAMING_SNAKE_CASE__ : Dict =floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : List[str] =image.cpu().permute(0 , 2 , 3 , 1 )[0] SCREAMING_SNAKE_CASE__ : Optional[int] =Image.fromarray(np.uinta(__lowercase ) ).convert('''RGB''' ).resize((2_56, 2_56) ) if str(__lowercase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE__ : str =torch.manual_seed(__lowercase ) else: SCREAMING_SNAKE_CASE__ : Dict =torch.Generator(device=__lowercase ).manual_seed(__lowercase ) SCREAMING_SNAKE_CASE__ : Tuple ={ '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def __magic_name__ ( self : Union[str, Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ : List[str] ='''cpu''' SCREAMING_SNAKE_CASE__ : Optional[int] =self.get_dummy_components() SCREAMING_SNAKE_CASE__ : str =self.pipeline_class(**__lowercase ) SCREAMING_SNAKE_CASE__ : int =pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) SCREAMING_SNAKE_CASE__ : Optional[Any] =pipe(**self.get_dummy_inputs(__lowercase ) ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =output.images SCREAMING_SNAKE_CASE__ : List[Any] =pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] SCREAMING_SNAKE_CASE__ : Optional[Any] =image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE__ : str =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE__ : Any =np.array( [0.6199778, 0.63984406, 0.46145785, 0.62944984, 0.5622215, 0.47306132, 0.47441456, 0.4607606, 0.48719263] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __magic_name__ ( self : Dict ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self : List[str] ) -> Any: SCREAMING_SNAKE_CASE__ : Tuple =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_img2img_frog.npy''' ) SCREAMING_SNAKE_CASE__ : Any =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) SCREAMING_SNAKE_CASE__ : Tuple ='''A red cartoon frog, 4k''' SCREAMING_SNAKE_CASE__ : List[Any] =KandinskyVaaPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) SCREAMING_SNAKE_CASE__ : Optional[int] =KandinskyVaaImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-decoder''' , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE__ : str =pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) SCREAMING_SNAKE_CASE__ : Dict =torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : str =pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() SCREAMING_SNAKE_CASE__ : Dict =pipeline( image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , strength=0.2 , output_type='''np''' , ) SCREAMING_SNAKE_CASE__ : Any =output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
152
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , __lowercase : int ) -> None: SCREAMING_SNAKE_CASE__ : List[Any] =value SCREAMING_SNAKE_CASE__ : Node | None =None SCREAMING_SNAKE_CASE__ : Node | None =None class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , __lowercase : Node ) -> None: SCREAMING_SNAKE_CASE__ : Any =tree def __magic_name__ ( self : str , __lowercase : Node | None ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : Union[str, Any] ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
152
1
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) class _lowerCamelCase( lowerCamelCase_ ): def __init__( self, *lowerCamelCase, **lowerCamelCase) -> None: """simple docstring""" warnings.warn( 'The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ChineseCLIPImageProcessor instead.', __snake_case, ) super().__init__(*__snake_case, **__snake_case)
370
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _lowerCamelCase( _a ): lowercase_ : Union[str, Any] = """char""" lowercase_ : Any = """bpe""" lowercase_ : Optional[int] = """wp""" SCREAMING_SNAKE_CASE : Optional[int] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _lowerCamelCase( _a ): lowercase_ : Any = ["""image_processor""", """char_tokenizer"""] lowercase_ : Tuple = """ViTImageProcessor""" lowercase_ : List[str] = """MgpstrTokenizer""" def __init__( self, lowerCamelCase=None, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : Any = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.', lowerCamelCase, ) _lowercase : str = kwargs.pop('feature_extractor') _lowercase : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.') if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.') _lowercase : List[Any] = tokenizer _lowercase : Tuple = AutoTokenizer.from_pretrained('gpt2') _lowercase : Tuple = AutoTokenizer.from_pretrained('bert-base-uncased') super().__init__(lowerCamelCase, lowerCamelCase) def __call__( self, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" if images is None and text is None: raise ValueError('You need to specify either an `images` or `text` input to process.') if images is not None: _lowercase : Optional[Any] = self.image_processor(lowerCamelCase, return_tensors=lowerCamelCase, **lowerCamelCase) if text is not None: _lowercase : Optional[int] = self.char_tokenizer(lowerCamelCase, return_tensors=lowerCamelCase, **lowerCamelCase) if text is None: return inputs elif images is None: return encodings else: _lowercase : Optional[int] = encodings['input_ids'] return inputs def UpperCamelCase ( self, lowerCamelCase) -> Any: """simple docstring""" _lowercase , _lowercase , _lowercase : Optional[int] = sequences _lowercase : str = char_preds.size(0) _lowercase , _lowercase : List[Any] = self._decode_helper(lowerCamelCase, 'char') _lowercase , _lowercase : str = self._decode_helper(lowerCamelCase, 'bpe') _lowercase , _lowercase : str = self._decode_helper(lowerCamelCase, 'wp') _lowercase : Dict = [] _lowercase : Any = [] for i in range(lowerCamelCase): _lowercase : Optional[int] = [char_scores[i], bpe_scores[i], wp_scores[i]] _lowercase : List[Any] = [char_strs[i], bpe_strs[i], wp_strs[i]] _lowercase : Union[str, Any] = scores.index(max(lowerCamelCase)) final_strs.append(strs[max_score_index]) final_scores.append(scores[max_score_index]) _lowercase : str = {} _lowercase : int = final_strs _lowercase : Optional[Any] = final_scores _lowercase : Tuple = char_strs _lowercase : Dict = bpe_strs _lowercase : Tuple = wp_strs return out def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> str: """simple docstring""" if format == DecodeType.CHARACTER: _lowercase : Optional[Any] = self.char_decode _lowercase : int = 1 _lowercase : int = '[s]' elif format == DecodeType.BPE: _lowercase : List[Any] = self.bpe_decode _lowercase : Union[str, Any] = 2 _lowercase : Any = '#' elif format == DecodeType.WORDPIECE: _lowercase : int = self.wp_decode _lowercase : Optional[Any] = 1_02 _lowercase : List[Any] = '[SEP]' else: raise ValueError(F'''Format {format} is not supported.''') _lowercase , _lowercase : Tuple = [], [] _lowercase : str = pred_logits.size(0) _lowercase : Tuple = pred_logits.size(1) _lowercase , _lowercase : Dict = pred_logits.topk(1, dim=-1, largest=lowerCamelCase, sorted=lowerCamelCase) _lowercase : List[str] = preds_index.view(-1, lowerCamelCase)[:, 1:] _lowercase : int = decoder(lowerCamelCase) _lowercase , _lowercase : Optional[Any] = torch.nn.functional.softmax(lowerCamelCase, dim=2).max(dim=2) _lowercase : Optional[Any] = preds_max_prob[:, 1:] for index in range(lowerCamelCase): _lowercase : List[str] = preds_str[index].find(lowerCamelCase) _lowercase : int = preds_str[index][:pred_eos] _lowercase : List[str] = preds_index[index].cpu().tolist() _lowercase : Optional[int] = pred_index.index(lowerCamelCase) if eos_token in pred_index else -1 _lowercase : int = preds_max_prob[index][: pred_eos_index + 1] _lowercase : Tuple = pred_max_prob.cumprod(dim=0)[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(lowerCamelCase) conf_scores.append(lowerCamelCase) return dec_strs, conf_scores def UpperCamelCase ( self, lowerCamelCase) -> Any: """simple docstring""" _lowercase : Dict = [seq.replace(' ', '') for seq in self.char_tokenizer.batch_decode(lowerCamelCase)] return decode_strs def UpperCamelCase ( self, lowerCamelCase) -> Union[str, Any]: """simple docstring""" return self.bpe_tokenizer.batch_decode(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = [seq.replace(' ', '') for seq in self.wp_tokenizer.batch_decode(lowerCamelCase)] return decode_strs
84
0
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class lowerCamelCase (A__ ): """simple docstring""" def __init__( self : Optional[int] , __magic_name__ : Tuple , __magic_name__ : Tuple ) -> List[Any]: SCREAMING_SNAKE_CASE_ = params SCREAMING_SNAKE_CASE_ = np.array(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = np.array([len(_lowerCAmelCase ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : List[str] , __magic_name__ : int ) -> List[str]: return (self.token_ids[index], self.lengths[index]) def __len__( self : str ) -> int: return len(self.lengths ) def __A ( self : Tuple ) -> int: assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def __A ( self : List[Any] ) -> Dict: SCREAMING_SNAKE_CASE_ = self.params.max_model_input_size SCREAMING_SNAKE_CASE_ = self.lengths > max_len logger.info(F'''Splitting {sum(_lowerCAmelCase )} too long sequences.''' ) def divide_chunks(__magic_name__ : Dict , __magic_name__ : List[str] ): return [l[i : i + n] for i in range(0 , len(_lowerCAmelCase ) , _lowerCAmelCase )] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] if self.params.mlm: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.params.special_tok_ids["cls_token"], self.params.special_tok_ids["sep_token"] else: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.params.special_tok_ids["bos_token"], self.params.special_tok_ids["eos_token"] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: SCREAMING_SNAKE_CASE_ = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: SCREAMING_SNAKE_CASE_ = np.insert(_lowerCAmelCase , 0 , _lowerCAmelCase ) if sub_s[-1] != sep_id: SCREAMING_SNAKE_CASE_ = np.insert(_lowerCAmelCase , len(_lowerCAmelCase ) , _lowerCAmelCase ) assert len(_lowerCAmelCase ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(_lowerCAmelCase ) new_tok_ids.extend(_lowerCAmelCase ) new_lengths.extend([len(_lowerCAmelCase ) for l in sub_seqs] ) SCREAMING_SNAKE_CASE_ = np.array(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = np.array(_lowerCAmelCase ) def __A ( self : str ) -> List[Any]: SCREAMING_SNAKE_CASE_ = len(self ) SCREAMING_SNAKE_CASE_ = self.lengths > 11 SCREAMING_SNAKE_CASE_ = self.token_ids[indices] SCREAMING_SNAKE_CASE_ = self.lengths[indices] SCREAMING_SNAKE_CASE_ = len(self ) logger.info(F'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def __A ( self : int ) -> Optional[Any]: if "unk_token" not in self.params.special_tok_ids: return else: SCREAMING_SNAKE_CASE_ = self.params.special_tok_ids["unk_token"] SCREAMING_SNAKE_CASE_ = len(self ) SCREAMING_SNAKE_CASE_ = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) SCREAMING_SNAKE_CASE_ = (unk_occs / self.lengths) < 0.5 SCREAMING_SNAKE_CASE_ = self.token_ids[indices] SCREAMING_SNAKE_CASE_ = self.lengths[indices] SCREAMING_SNAKE_CASE_ = len(self ) logger.info(F'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def __A ( self : Dict ) -> Optional[Any]: if not self.params.is_master: return logger.info(F'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def __A ( self : Dict , __magic_name__ : List[str] ) -> Tuple: SCREAMING_SNAKE_CASE_ = [t[0] for t in batch] SCREAMING_SNAKE_CASE_ = [t[1] for t in batch] assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) # Max for paddings SCREAMING_SNAKE_CASE_ = max(_lowerCAmelCase ) # Pad token ids if self.params.mlm: SCREAMING_SNAKE_CASE_ = self.params.special_tok_ids["pad_token"] else: SCREAMING_SNAKE_CASE_ = self.params.special_tok_ids["unk_token"] SCREAMING_SNAKE_CASE_ = [list(t.astype(_lowerCAmelCase ) ) + [pad_idx] * (max_seq_len_ - len(_lowerCAmelCase )) for t in token_ids] assert len(tk_ ) == len(_lowerCAmelCase ) assert all(len(_lowerCAmelCase ) == max_seq_len_ for t in tk_ ) SCREAMING_SNAKE_CASE_ = torch.tensor(tk_ ) # (bs, max_seq_len_) SCREAMING_SNAKE_CASE_ = torch.tensor(_lowerCAmelCase ) # (bs) return tk_t, lg_t
118
'''simple docstring''' def __a ( UpperCAmelCase , UpperCAmelCase ) ->int: """simple docstring""" return int((input_a, input_a).count(1 ) != 0 ) def __a ( ) ->None: """simple docstring""" assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
258
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class __magic_name__ ( unittest.TestCase): UpperCamelCase__ = MODEL_FOR_CAUSAL_LM_MAPPING UpperCamelCase__ = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : List[str] = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""pt""" ) # Using `do_sample=False` to force deterministic output lowercase_ : Union[str, Any] = text_generator("""This is a test""" , do_sample=lowercase_ ) self.assertEqual( lowercase_ , [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ] , ) lowercase_ : Optional[Any] = text_generator(["""This is a test""", """This is a second test"""] ) self.assertEqual( lowercase_ , [ [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ], [ { """generated_text""": ( """This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy""" """ oscope. oscope. FiliFili@@""" ) } ], ] , ) lowercase_ : Dict = text_generator("""This is a test""" , do_sample=lowercase_ , num_return_sequences=2 , return_tensors=lowercase_ ) self.assertEqual( lowercase_ , [ {"""generated_token_ids""": ANY(lowercase_ )}, {"""generated_token_ids""": ANY(lowercase_ )}, ] , ) lowercase_ : Any = text_generator.model.config.eos_token_id lowercase_ : List[Any] = """<pad>""" lowercase_ : Dict = text_generator( ["""This is a test""", """This is a second test"""] , do_sample=lowercase_ , num_return_sequences=2 , batch_size=2 , return_tensors=lowercase_ , ) self.assertEqual( lowercase_ , [ [ {"""generated_token_ids""": ANY(lowercase_ )}, {"""generated_token_ids""": ANY(lowercase_ )}, ], [ {"""generated_token_ids""": ANY(lowercase_ )}, {"""generated_token_ids""": ANY(lowercase_ )}, ], ] , ) @require_tf def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : List[str] = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""tf""" ) # Using `do_sample=False` to force deterministic output lowercase_ : Tuple = text_generator("""This is a test""" , do_sample=lowercase_ ) self.assertEqual( lowercase_ , [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ] , ) lowercase_ : Optional[Any] = text_generator(["""This is a test""", """This is a second test"""] , do_sample=lowercase_ ) self.assertEqual( lowercase_ , [ [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ], [ { """generated_text""": ( """This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes""" """ Cannes 閲閲Cannes Cannes Cannes 攵 please,""" ) } ], ] , ) def SCREAMING_SNAKE_CASE_ ( self : Dict , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : int ): lowercase_ : Tuple = TextGenerationPipeline(model=lowercase_ , tokenizer=lowercase_ ) return text_generator, ["This is a test", "Another test"] def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : Optional[int] = """Hello I believe in""" lowercase_ : Union[str, Any] = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) lowercase_ : Union[str, Any] = text_generator(lowercase_ ) self.assertEqual( lowercase_ , [{"""generated_text""": """Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe"""}] , ) lowercase_ : int = text_generator(lowercase_ , stop_sequence=""" fe""" ) self.assertEqual(lowercase_ , [{"""generated_text""": """Hello I believe in fe"""}] ) def SCREAMING_SNAKE_CASE_ ( self : str , lowercase_ : Any , lowercase_ : Dict ): lowercase_ : Any = text_generator.model lowercase_ : Tuple = text_generator.tokenizer lowercase_ : str = text_generator("""This is a test""" ) self.assertEqual(lowercase_ , [{"""generated_text""": ANY(lowercase_ )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) lowercase_ : str = text_generator("""This is a test""" , return_full_text=lowercase_ ) self.assertEqual(lowercase_ , [{"""generated_text""": ANY(lowercase_ )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) lowercase_ : Tuple = pipeline(task="""text-generation""" , model=lowercase_ , tokenizer=lowercase_ , return_full_text=lowercase_ ) lowercase_ : Tuple = text_generator("""This is a test""" ) self.assertEqual(lowercase_ , [{"""generated_text""": ANY(lowercase_ )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) lowercase_ : Optional[Any] = text_generator("""This is a test""" , return_full_text=lowercase_ ) self.assertEqual(lowercase_ , [{"""generated_text""": ANY(lowercase_ )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) lowercase_ : int = text_generator(["""This is great !""", """Something else"""] , num_return_sequences=2 , do_sample=lowercase_ ) self.assertEqual( lowercase_ , [ [{"""generated_text""": ANY(lowercase_ )}, {"""generated_text""": ANY(lowercase_ )}], [{"""generated_text""": ANY(lowercase_ )}, {"""generated_text""": ANY(lowercase_ )}], ] , ) if text_generator.tokenizer.pad_token is not None: lowercase_ : Optional[Any] = text_generator( ["""This is great !""", """Something else"""] , num_return_sequences=2 , batch_size=2 , do_sample=lowercase_ ) self.assertEqual( lowercase_ , [ [{"""generated_text""": ANY(lowercase_ )}, {"""generated_text""": ANY(lowercase_ )}], [{"""generated_text""": ANY(lowercase_ )}, {"""generated_text""": ANY(lowercase_ )}], ] , ) with self.assertRaises(lowercase_ ): lowercase_ : Dict = text_generator("""test""" , return_full_text=lowercase_ , return_text=lowercase_ ) with self.assertRaises(lowercase_ ): lowercase_ : Union[str, Any] = text_generator("""test""" , return_full_text=lowercase_ , return_tensors=lowercase_ ) with self.assertRaises(lowercase_ ): lowercase_ : Optional[int] = text_generator("""test""" , return_text=lowercase_ , return_tensors=lowercase_ ) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): lowercase_ : Tuple = text_generator("""""" ) self.assertEqual(lowercase_ , [{"""generated_text""": ANY(lowercase_ )}] ) else: with self.assertRaises((ValueError, AssertionError) ): lowercase_ : List[Any] = text_generator("""""" ) if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. lowercase_ : Tuple = ["""RwkvForCausalLM""", """XGLMForCausalLM""", """GPTNeoXForCausalLM"""] if ( tokenizer.model_max_length < 10000 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError) ): text_generator("""This is a test""" * 500 , max_new_tokens=20 ) lowercase_ : Union[str, Any] = text_generator("""This is a test""" * 500 , handle_long_generation="""hole""" , max_new_tokens=20 ) # Hole strategy cannot work with self.assertRaises(lowercase_ ): text_generator( """This is a test""" * 500 , handle_long_generation="""hole""" , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def SCREAMING_SNAKE_CASE_ ( self : Any ): import torch # Classic `model_kwargs` lowercase_ : Any = pipeline( model="""hf-internal-testing/tiny-random-bloom""" , model_kwargs={"""device_map""": """auto""", """torch_dtype""": torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowercase_ : int = pipe("""This is a test""" ) self.assertEqual( lowercase_ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) lowercase_ : Tuple = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.bfloataa ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowercase_ : Tuple = pipe("""This is a test""" ) self.assertEqual( lowercase_ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 lowercase_ : Any = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa ) lowercase_ : Optional[Any] = pipe("""This is a test""" ) self.assertEqual( lowercase_ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) @require_torch @require_torch_gpu def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): import torch lowercase_ : Optional[Any] = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device=0 , torch_dtype=torch.floataa ) pipe("""This is a test""" ) @require_torch @require_accelerate @require_torch_gpu def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): import torch lowercase_ : Any = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.floataa ) pipe("""This is a test""" , do_sample=lowercase_ , top_p=0.5 ) def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Union[str, Any] = """Hello world""" lowercase_ : Any = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) if text_generator.model.framework == "tf": lowercase_ : int = logging.get_logger("""transformers.generation.tf_utils""" ) else: lowercase_ : Union[str, Any] = logging.get_logger("""transformers.generation.utils""" ) lowercase_ : List[Any] = """Both `max_new_tokens`""" # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(lowercase_ ) as cl: lowercase_ : List[Any] = text_generator(lowercase_ , max_length=10 , max_new_tokens=1 ) self.assertIn(lowercase_ , cl.out ) # The user only sets one -> no warning with CaptureLogger(lowercase_ ) as cl: lowercase_ : List[str] = text_generator(lowercase_ , max_new_tokens=1 ) self.assertNotIn(lowercase_ , cl.out ) with CaptureLogger(lowercase_ ) as cl: lowercase_ : Optional[Any] = text_generator(lowercase_ , max_length=10 ) self.assertNotIn(lowercase_ , cl.out )
21
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline _lowercase : Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class __magic_name__ ( _UpperCAmelCase): def __init__( self : Union[str, Any] , lowercase_ : Optional[int] , lowercase_ : str ): super().__init__() self.register_modules(unet=lowercase_ , scheduler=lowercase_ ) @torch.no_grad() def __call__( self : List[str] , lowercase_ : int = 1 , lowercase_ : int = 100 , lowercase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowercase_ : Optional[float] = None , lowercase_ : bool = True , ): if audio_length_in_s is None: lowercase_ : List[Any] = self.unet.config.sample_size / self.unet.config.sample_rate lowercase_ : Dict = audio_length_in_s * self.unet.config.sample_rate lowercase_ : Any = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( f'''{audio_length_in_s} is too small. Make sure it\'s bigger or equal to''' f''' {3 * down_scale_factor / self.unet.config.sample_rate}.''' ) lowercase_ : List[Any] = int(lowercase_ ) if sample_size % down_scale_factor != 0: lowercase_ : int = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( f'''{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled''' f''' by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising''' """ process.""" ) lowercase_ : Any = int(lowercase_ ) lowercase_ : List[str] = next(iter(self.unet.parameters() ) ).dtype lowercase_ : List[str] = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(lowercase_ )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) lowercase_ : Any = randn_tensor(lowercase_ , generator=lowercase_ , device=self.device , dtype=lowercase_ ) # set step values self.scheduler.set_timesteps(lowercase_ , device=audio.device ) lowercase_ : Optional[Any] = self.scheduler.timesteps.to(lowercase_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowercase_ : Dict = self.unet(lowercase_ , lowercase_ ).sample # 2. compute previous image: x_t -> t_t-1 lowercase_ : List[str] = self.scheduler.step(lowercase_ , lowercase_ , lowercase_ ).prev_sample lowercase_ : str = audio.clamp(-1 , 1 ).float().cpu().numpy() lowercase_ : Union[str, Any] = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=lowercase_ )
21
1
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer UpperCAmelCase_ : int = '''bart''' UpperCAmelCase_ : int = True @st.cache(allow_output_mutation=__magic_name__ ) def SCREAMING_SNAKE_CASE_ ( ) -> Optional[int]: """simple docstring""" if LOAD_DENSE_INDEX: UpperCamelCase :Optional[Any] = AutoTokenizer.from_pretrained("""yjernite/retribert-base-uncased""" ) UpperCamelCase :Union[str, Any] = AutoModel.from_pretrained("""yjernite/retribert-base-uncased""" ).to("""cuda:0""" ) UpperCamelCase :Optional[int] = qar_model.eval() else: UpperCamelCase , UpperCamelCase :Optional[int] = (None, None) if MODEL_TYPE == "bart": UpperCamelCase :Any = AutoTokenizer.from_pretrained("""yjernite/bart_eli5""" ) UpperCamelCase :Union[str, Any] = AutoModelForSeqaSeqLM.from_pretrained("""yjernite/bart_eli5""" ).to("""cuda:0""" ) UpperCamelCase :Optional[int] = torch.load("""seq2seq_models/eli5_bart_model_blm_2.pth""" ) sas_model.load_state_dict(save_dict["""model"""] ) UpperCamelCase :Optional[Any] = sas_model.eval() else: UpperCamelCase , UpperCamelCase :Dict = make_qa_sas_model( model_name="""t5-small""" , from_file="""seq2seq_models/eli5_t5_model_1024_4.pth""" , device="""cuda:0""" ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=__magic_name__ ) def SCREAMING_SNAKE_CASE_ ( ) -> Union[str, Any]: """simple docstring""" if LOAD_DENSE_INDEX: UpperCamelCase :Tuple = faiss.StandardGpuResources() UpperCamelCase :Dict = datasets.load_dataset(path="""wiki_snippets""" , name="""wiki40b_en_100_0""" )["""train"""] UpperCamelCase :str = np.memmap( """wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat""" , dtype="""float32""" , mode="""r""" , shape=(wikiaab_passages.num_rows, 128) , ) UpperCamelCase :Optional[Any] = faiss.IndexFlatIP(128 ) UpperCamelCase :Optional[int] = faiss.index_cpu_to_gpu(__magic_name__ , 1 , __magic_name__ ) wikiaab_gpu_index_flat.add(__magic_name__ ) # TODO fix for larger GPU else: UpperCamelCase , UpperCamelCase :Optional[Any] = (None, None) UpperCamelCase :List[str] = Elasticsearch([{"""host""": """localhost""", """port""": """9200"""}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=__magic_name__ ) def SCREAMING_SNAKE_CASE_ ( ) -> str: """simple docstring""" UpperCamelCase :Dict = datasets.load_dataset("""eli5""" , name="""LFQA_reddit""" ) UpperCamelCase :List[Any] = elia["""train_eli5"""] UpperCamelCase :Optional[Any] = np.memmap( """eli5_questions_reps.dat""" , dtype="""float32""" , mode="""r""" , shape=(elia_train.num_rows, 128) ) UpperCamelCase :Dict = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(__magic_name__ ) return (elia_train, eli5_train_q_index) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = load_indexes() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = load_models() UpperCAmelCase_ , UpperCAmelCase_ : str = load_train_data() def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int , __magic_name__ : Any=10 ) -> Any: """simple docstring""" UpperCamelCase :List[str] = embed_questions_for_retrieval([question] , __magic_name__ , __magic_name__ ) UpperCamelCase , UpperCamelCase :int = eli5_train_q_index.search(__magic_name__ , __magic_name__ ) UpperCamelCase :Any = [elia_train[int(__magic_name__ )] for i in I[0]] return nn_examples def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Union[str, Any] , __magic_name__ : List[str]="wiki40b" , __magic_name__ : str="dense" , __magic_name__ : Tuple=10 ) -> List[str]: """simple docstring""" if source == "none": UpperCamelCase , UpperCamelCase :Dict = (""" <P> """.join(["""""" for _ in range(11 )] ).strip(), []) else: if method == "dense": UpperCamelCase , UpperCamelCase :List[Any] = query_qa_dense_index( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: UpperCamelCase , UpperCamelCase :List[Any] = query_es_index( __magic_name__ , __magic_name__ , index_name="""english_wiki40b_snippets_100w""" , n_results=__magic_name__ , ) UpperCamelCase :str = [ (res["""article_title"""], res["""section_title"""].strip(), res["""score"""], res["""passage_text"""]) for res in hit_lst ] UpperCamelCase :Tuple = """question: {} context: {}""".format(__magic_name__ , __magic_name__ ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda __magic_name__ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda __magic_name__ : None), } ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : List[str] , __magic_name__ : Union[str, Any] , __magic_name__ : List[str] , __magic_name__ : Optional[Any]=64 , __magic_name__ : int=256 , __magic_name__ : Dict=False , __magic_name__ : str=2 , __magic_name__ : str=0.95 , __magic_name__ : Dict=0.8 ) -> Dict: """simple docstring""" with torch.no_grad(): UpperCamelCase :Optional[Any] = qa_sas_generate( __magic_name__ , __magic_name__ , __magic_name__ , num_answers=1 , num_beams=__magic_name__ , min_len=__magic_name__ , max_len=__magic_name__ , do_sample=__magic_name__ , temp=__magic_name__ , top_p=__magic_name__ , top_k=__magic_name__ , max_input_length=1024 , device="""cuda:0""" , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar UpperCAmelCase_ : List[Any] = '''<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>''' UpperCAmelCase_ : Union[str, Any] = ''' <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class="img-container"> <!-- Inline parent element --> %s </span> </body> </html> ''' % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia UpperCAmelCase_ : List[str] = ''' This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. ''' st.sidebar.markdown(description, unsafe_allow_html=True) UpperCAmelCase_ : Dict = [ '''Answer the question''', '''View the retrieved document only''', '''View the most similar ELI5 question and answer''', '''Show me everything, please!''', ] UpperCAmelCase_ : Tuple = st.sidebar.checkbox('''Demo options''') if demo_options: UpperCAmelCase_ : str = st.sidebar.selectbox( '''''', action_list, index=3, ) UpperCAmelCase_ : str = action_list.index(action_st) UpperCAmelCase_ : int = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) UpperCAmelCase_ : str = show_type == '''Show full text of passages''' else: UpperCAmelCase_ : str = 3 UpperCAmelCase_ : List[Any] = True UpperCAmelCase_ : Optional[Any] = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: UpperCAmelCase_ : Any = ''' ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. ''' st.sidebar.markdown(retriever_info) UpperCAmelCase_ : int = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) UpperCAmelCase_ : List[Any] = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: UpperCAmelCase_ : Optional[Any] = '''wiki40b''' UpperCAmelCase_ : Any = '''dense''' UpperCAmelCase_ : int = '''beam''' UpperCAmelCase_ : Optional[int] = 2 UpperCAmelCase_ : Optional[Any] = 64 UpperCAmelCase_ : str = 2_56 UpperCAmelCase_ : Any = None UpperCAmelCase_ : List[Any] = None UpperCAmelCase_ : str = st.sidebar.checkbox('''Generation options''') if generate_options: UpperCAmelCase_ : Optional[Any] = ''' ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder\'s output probabilities. ''' st.sidebar.markdown(generate_info) UpperCAmelCase_ : Optional[Any] = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) UpperCAmelCase_ : int = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=2_56, value=64, step=8, format=None, key=None ) UpperCAmelCase_ : Optional[Any] = st.sidebar.slider( '''Maximum generation length''', min_value=64, max_value=5_12, value=2_56, step=16, format=None, key=None ) if sampled == "beam": UpperCAmelCase_ : Optional[Any] = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: UpperCAmelCase_ : int = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) UpperCAmelCase_ : Tuple = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) UpperCAmelCase_ : Optional[int] = None # start main text UpperCAmelCase_ : Optional[int] = [ '''<MY QUESTION>''', '''How do people make chocolate?''', '''Why do we get a fever when we are sick?''', '''How can different animals perceive different colors?''', '''What is natural language processing?''', '''What\'s the best way to treat a sunburn?''', '''What exactly are vitamins ?''', '''How does nuclear energy provide electricity?''', '''What\'s the difference between viruses and bacteria?''', '''Why are flutes classified as woodwinds when most of them are made out of metal ?''', '''Why do people like drinking coffee even though it tastes so bad?''', '''What happens when wine ages? How does it make the wine taste better?''', '''If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?''', '''How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?''', '''How does New Zealand have so many large bird predators?''', ] UpperCAmelCase_ : List[Any] = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": UpperCAmelCase_ : Union[str, Any] = st.text_input('''Enter your question here:''', '''''') else: UpperCAmelCase_ : Optional[Any] = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": UpperCAmelCase_ , UpperCAmelCase_ : Dict = make_support(question, source=wiki_source, method='''dense''', n_results=10) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = make_support(question, source=wiki_source, method='''sparse''', n_results=10) UpperCAmelCase_ : Tuple = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] UpperCAmelCase_ : Any = support_list[:10] UpperCAmelCase_ : Optional[Any] = '''<P> ''' + ''' <P> '''.join([res[-1] for res in support_list]) else: UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: UpperCAmelCase_ , UpperCAmelCase_ : str = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): UpperCAmelCase_ : Any = '''https://en.wikipedia.org/wiki/{}'''.format(res[0].replace(''' ''', '''_''')) UpperCAmelCase_ : List[Any] = res[1].strip() if sec_titles == "": UpperCAmelCase_ : Union[str, Any] = '''[{}]({})'''.format(res[0], wiki_url) else: UpperCAmelCase_ : str = sec_titles.split(''' & ''') UpperCAmelCase_ : str = ''' & '''.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style="font-family:arial; font-size:10pt;">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: UpperCAmelCase_ : Any = find_nearest_training(question) UpperCAmelCase_ : Optional[int] = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) UpperCAmelCase_ : Dict = [ '''{}. {}'''.format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) UpperCAmelCase_ : int = ''' --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* ''' st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
38
import random def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" snake_case = a[left_index] snake_case = left_index + 1 for j in range(left_index + 1 ,UpperCamelCase_ ): if a[j] < pivot: snake_case , snake_case = a[i], a[j] i += 1 snake_case , snake_case = a[i - 1], a[left_index] return i - 1 def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" if left < right: snake_case = random.randint(UpperCamelCase_ ,right - 1 ) snake_case , snake_case = ( a[left], a[pivot], ) # switches the pivot with the left most bound snake_case = partition(UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ) quick_sort_random( UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ) # recursive quicksort to the left of the pivot point quick_sort_random( UpperCamelCase_ ,pivot_index + 1 ,UpperCamelCase_ ) # recursive quicksort to the right of the pivot point def UpperCAmelCase__ (): """simple docstring""" snake_case = input('''Enter numbers separated by a comma:\n''' ).strip() snake_case = [int(UpperCamelCase_ ) for item in user_input.split(''',''' )] quick_sort_random(UpperCamelCase_ ,0 ,len(UpperCamelCase_ ) ) print(UpperCamelCase_ ) if __name__ == "__main__": main()
127
0
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : str ): '''simple docstring''' lowerCamelCase_ = len(lowerCAmelCase__ ) lowerCamelCase_ = len(lowerCAmelCase__ ) lowerCamelCase_ = ( first_str_length if first_str_length > second_str_length else second_str_length ) lowerCamelCase_ = [] for char_count in range(lowerCAmelCase__ ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(lowerCAmelCase__ ) if __name__ == "__main__": print(alternative_string_arrange("AB", "XYZ"), end=" ")
360
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int ): '''simple docstring''' return [sentence[i : i + ngram_size] for i in range(len(lowercase ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
208
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
61
"""simple docstring""" from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __A = Lock() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[Any]: """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(_SCREAMING_SNAKE_CASE ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowerCAmelCase__ :Any = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowerCAmelCase__ :Tuple = min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(_SCREAMING_SNAKE_CASE ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowerCAmelCase__ :Optional[int] = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowerCAmelCase__ :Optional[int] = max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # after all swaps are performed, send the values back to main result_pipe[1].send(_SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" lowerCAmelCase__ :str = [] lowerCAmelCase__ :Optional[Any] = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowerCAmelCase__ :List[str] = Pipe() lowerCAmelCase__ :List[Any] = Pipe() process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowerCAmelCase__ :Dict = temp_rs lowerCAmelCase__ :Optional[Any] = temp_rr for i in range(1 , len(_SCREAMING_SNAKE_CASE ) - 1 ): lowerCAmelCase__ :Union[str, Any] = Pipe() lowerCAmelCase__ :List[str] = Pipe() process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowerCAmelCase__ :Union[str, Any] = temp_rs lowerCAmelCase__ :Any = temp_rr process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=( len(_SCREAMING_SNAKE_CASE ) - 1, arr[len(_SCREAMING_SNAKE_CASE ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(_SCREAMING_SNAKE_CASE ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(_SCREAMING_SNAKE_CASE ) ): lowerCAmelCase__ :str = result_pipe[p][0].recv() process_array_[p].join() return arr def __A () ->List[Any]: """simple docstring""" lowerCAmelCase__ :Union[str, Any] = list(range(10 , 0 , -1 ) ) print('Initial List' ) print(*_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[str] = odd_even_transposition(_SCREAMING_SNAKE_CASE ) print('Sorted List\n' ) print(*_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
293
0
"""simple docstring""" import math def lowerCamelCase_ (UpperCamelCase__ : int = 100 ): _UpperCAmelCase : Any = sum(i * i for i in range(1 , n + 1 ) ) _UpperCAmelCase : Optional[int] = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(f"{solution() = }")
354
"""simple docstring""" import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _lowerCAmelCase :int = get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' a__ ='''all_checks''' a__ ='''basic_checks''' a__ ='''no_checks''' class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' def lowerCamelCase_ (UpperCamelCase__ : Optional[dict] , UpperCamelCase__ : dict , UpperCamelCase__ : Tuple=None ): if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) ) if len(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) > 0: raise UnexpectedDownloadedFile(str(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) ) _UpperCAmelCase : Optional[Any] = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase : str = ''' for ''' + verification_name if verification_name is not None else '''''' if len(UpperCamelCase__ ) > 0: raise NonMatchingChecksumError( F'Checksums didn\'t match{for_verification_name}:\n' F'{bad_urls}\n' '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' class _UpperCAmelCase ( a ): '''simple docstring''' def lowerCamelCase_ (UpperCamelCase__ : Optional[dict] , UpperCamelCase__ : dict ): if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) > 0: raise ExpectedMoreSplits(str(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) ) if len(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) > 0: raise UnexpectedSplits(str(set(UpperCamelCase__ ) - set(UpperCamelCase__ ) ) ) _UpperCAmelCase : Dict = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(UpperCamelCase__ ) > 0: raise NonMatchingSplitsSizesError(str(UpperCamelCase__ ) ) logger.info('''All the splits matched successfully.''' ) def lowerCamelCase_ (UpperCamelCase__ : str , UpperCamelCase__ : bool = True ): if record_checksum: _UpperCAmelCase : Any = shaaaa() with open(UpperCamelCase__ , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'''''' ): m.update(UpperCamelCase__ ) _UpperCAmelCase : int = m.hexdigest() else: _UpperCAmelCase : Union[str, Any] = None return {"num_bytes": os.path.getsize(UpperCamelCase__ ), "checksum": checksum} def lowerCamelCase_ (UpperCamelCase__ : List[str] ): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
68
0
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __UpperCAmelCase : int = logging.get_logger(__name__) __UpperCAmelCase : List[str] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase : Any = { "vocab_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/vocab.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/vocab.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/vocab.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/vocab.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/vocab.json", }, "merges_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/merges.txt", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/merges.txt", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/merges.txt", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/merges.txt", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/merges.txt", }, "tokenizer_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/tokenizer.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/tokenizer.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/tokenizer.json", }, } __UpperCAmelCase : Optional[int] = { "gpt2": 1_024, "gpt2-medium": 1_024, "gpt2-large": 1_024, "gpt2-xl": 1_024, "distilgpt2": 1_024, } class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ["""input_ids""", """attention_mask"""] lowerCAmelCase__ = GPTaTokenizer def __init__( self : List[Any] , A : Dict=None , A : str=None , A : Optional[int]=None , A : List[str]="<|endoftext|>" , A : Union[str, Any]="<|endoftext|>" , A : Union[str, Any]="<|endoftext|>" , A : str=False , **A : Dict , ): super().__init__( _a , _a , tokenizer_file=_a , unk_token=_a , bos_token=_a , eos_token=_a , add_prefix_space=_a , **_a , ) __snake_case: Dict = kwargs.pop("""add_bos_token""" , _a ) __snake_case: Union[str, Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , _a ) != add_prefix_space: __snake_case: Dict = getattr(_a , pre_tok_state.pop("""type""" ) ) __snake_case: Tuple = add_prefix_space __snake_case: List[str] = pre_tok_class(**_a ) __snake_case: str = add_prefix_space def UpperCAmelCase__ ( self : int , *A : Dict , **A : Tuple ): __snake_case: str = kwargs.get("""is_split_into_words""" , _a ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*_a , **_a ) def UpperCAmelCase__ ( self : Tuple , *A : int , **A : Optional[int] ): __snake_case: Union[str, Any] = kwargs.get("""is_split_into_words""" , _a ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*_a , **_a ) def UpperCAmelCase__ ( self : List[str] , A : List[Any] , A : Optional[int] = None ): __snake_case: List[Any] = self._tokenizer.model.save(_a , name=_a ) return tuple(_a ) def UpperCAmelCase__ ( self : int , A : List[str] ): __snake_case: List[str] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_a , add_special_tokens=_a ) + [self.eos_token_id] ) if len(_a ) > self.model_max_length: __snake_case: str = input_ids[-self.model_max_length :] return input_ids
369
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __snake_case ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' pass @nightly @require_onnxruntime @require_torch_gpu class __snake_case ( unittest.TestCase ): '''simple docstring''' @property def UpperCAmelCase__ ( self : Dict ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCAmelCase__ ( self : Dict ): __snake_case: Optional[int] = ort.SessionOptions() __snake_case: List[Any] = False return options def UpperCAmelCase__ ( self : Optional[Any] ): __snake_case: Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __snake_case: Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __snake_case: List[str] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , safety_checker=A , feature_extractor=A , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A ) __snake_case: int = """A red cat sitting on a park bench""" __snake_case: Any = np.random.RandomState(0 ) __snake_case: Optional[Any] = pipe( prompt=A , image=A , mask_image=A , guidance_scale=7.5 , num_inference_steps=10 , generator=A , output_type="""np""" , ) __snake_case: List[Any] = output.images __snake_case: str = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __snake_case: Any = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCAmelCase__ ( self : Optional[Any] ): __snake_case: Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __snake_case: Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __snake_case: Optional[int] = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-inpainting""" , subfolder="""scheduler""" , revision="""onnx""" ) __snake_case: List[Any] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , scheduler=A , safety_checker=A , feature_extractor=A , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A ) __snake_case: Optional[int] = """A red cat sitting on a park bench""" __snake_case: Dict = np.random.RandomState(0 ) __snake_case: Optional[Any] = pipe( prompt=A , image=A , mask_image=A , guidance_scale=7.5 , num_inference_steps=20 , generator=A , output_type="""np""" , ) __snake_case: List[str] = output.images __snake_case: str = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __snake_case: Union[str, Any] = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
293
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ : Optional[Any] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase :List[str] = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: UpperCamelCase :Any = [144, 192, 240] UpperCamelCase :Any = [16, 32, 64, 96, 128, 160, 640] elif "mobilevit_xs" in mobilevit_name: UpperCamelCase :str = [96, 120, 144] UpperCamelCase :Tuple = [16, 32, 48, 64, 80, 96, 384] elif "mobilevit_xxs" in mobilevit_name: UpperCamelCase :int = [64, 80, 96] UpperCamelCase :List[Any] = [16, 16, 24, 48, 64, 80, 320] UpperCamelCase :Tuple = 0.05 UpperCamelCase :List[Any] = 2.0 if mobilevit_name.startswith("""deeplabv3_""" ): UpperCamelCase :str = 512 UpperCamelCase :int = 16 UpperCamelCase :List[str] = 21 UpperCamelCase :Tuple = """pascal-voc-id2label.json""" else: UpperCamelCase :Optional[int] = 1000 UpperCamelCase :Any = """imagenet-1k-id2label.json""" UpperCamelCase :Tuple = """huggingface/label-files""" UpperCamelCase :Tuple = json.load(open(hf_hub_download(__magic_name__ , __magic_name__ , repo_type="""dataset""" ) , """r""" ) ) UpperCamelCase :Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} UpperCamelCase :List[str] = idalabel UpperCamelCase :Dict = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[int] , __magic_name__ : List[str]=False ) -> Optional[Any]: """simple docstring""" for i in range(1 , 6 ): if f"""layer_{i}.""" in name: UpperCamelCase :Optional[int] = name.replace(f"""layer_{i}.""" , f"""encoder.layer.{i - 1}.""" ) if "conv_1." in name: UpperCamelCase :Tuple = name.replace("""conv_1.""" , """conv_stem.""" ) if ".block." in name: UpperCamelCase :Any = name.replace(""".block.""" , """.""" ) if "exp_1x1" in name: UpperCamelCase :str = name.replace("""exp_1x1""" , """expand_1x1""" ) if "red_1x1" in name: UpperCamelCase :Any = name.replace("""red_1x1""" , """reduce_1x1""" ) if ".local_rep.conv_3x3." in name: UpperCamelCase :Optional[Any] = name.replace(""".local_rep.conv_3x3.""" , """.conv_kxk.""" ) if ".local_rep.conv_1x1." in name: UpperCamelCase :Optional[Any] = name.replace(""".local_rep.conv_1x1.""" , """.conv_1x1.""" ) if ".norm." in name: UpperCamelCase :List[str] = name.replace(""".norm.""" , """.normalization.""" ) if ".conv." in name: UpperCamelCase :Dict = name.replace(""".conv.""" , """.convolution.""" ) if ".conv_proj." in name: UpperCamelCase :Any = name.replace(""".conv_proj.""" , """.conv_projection.""" ) for i in range(0 , 2 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: UpperCamelCase :Any = name.replace(f""".{i}.{j}.""" , f""".{i}.layer.{j}.""" ) for i in range(2 , 6 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: UpperCamelCase :str = name.replace(f""".{i}.{j}.""" , f""".{i}.""" ) if "expand_1x1" in name: UpperCamelCase :Any = name.replace("""expand_1x1""" , """downsampling_layer.expand_1x1""" ) if "conv_3x3" in name: UpperCamelCase :Any = name.replace("""conv_3x3""" , """downsampling_layer.conv_3x3""" ) if "reduce_1x1" in name: UpperCamelCase :List[str] = name.replace("""reduce_1x1""" , """downsampling_layer.reduce_1x1""" ) for i in range(2 , 5 ): if f""".global_rep.{i}.weight""" in name: UpperCamelCase :Tuple = name.replace(f""".global_rep.{i}.weight""" , """.layernorm.weight""" ) if f""".global_rep.{i}.bias""" in name: UpperCamelCase :str = name.replace(f""".global_rep.{i}.bias""" , """.layernorm.bias""" ) if ".global_rep." in name: UpperCamelCase :int = name.replace(""".global_rep.""" , """.transformer.""" ) if ".pre_norm_mha.0." in name: UpperCamelCase :Union[str, Any] = name.replace(""".pre_norm_mha.0.""" , """.layernorm_before.""" ) if ".pre_norm_mha.1.out_proj." in name: UpperCamelCase :Union[str, Any] = name.replace(""".pre_norm_mha.1.out_proj.""" , """.attention.output.dense.""" ) if ".pre_norm_ffn.0." in name: UpperCamelCase :Union[str, Any] = name.replace(""".pre_norm_ffn.0.""" , """.layernorm_after.""" ) if ".pre_norm_ffn.1." in name: UpperCamelCase :Any = name.replace(""".pre_norm_ffn.1.""" , """.intermediate.dense.""" ) if ".pre_norm_ffn.4." in name: UpperCamelCase :Any = name.replace(""".pre_norm_ffn.4.""" , """.output.dense.""" ) if ".transformer." in name: UpperCamelCase :Dict = name.replace(""".transformer.""" , """.transformer.layer.""" ) if ".aspp_layer." in name: UpperCamelCase :Optional[Any] = name.replace(""".aspp_layer.""" , """.""" ) if ".aspp_pool." in name: UpperCamelCase :Optional[Any] = name.replace(""".aspp_pool.""" , """.""" ) if "seg_head." in name: UpperCamelCase :str = name.replace("""seg_head.""" , """segmentation_head.""" ) if "segmentation_head.classifier.classifier." in name: UpperCamelCase :Dict = name.replace("""segmentation_head.classifier.classifier.""" , """segmentation_head.classifier.""" ) if "classifier.fc." in name: UpperCamelCase :Optional[int] = name.replace("""classifier.fc.""" , """classifier.""" ) elif (not base_model) and ("segmentation_head." not in name): UpperCamelCase :Any = """mobilevit.""" + name return name def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str , __magic_name__ : int , __magic_name__ : str=False ) -> Optional[Any]: """simple docstring""" if base_model: UpperCamelCase :Tuple = """""" else: UpperCamelCase :str = """mobilevit.""" for key in orig_state_dict.copy().keys(): UpperCamelCase :Any = orig_state_dict.pop(__magic_name__ ) if key[:8] == "encoder.": UpperCamelCase :Tuple = key[8:] if "qkv" in key: UpperCamelCase :Optional[int] = key.split(""".""" ) UpperCamelCase :Tuple = int(key_split[0][6:] ) - 1 UpperCamelCase :Dict = int(key_split[3] ) UpperCamelCase :Dict = model.get_submodule(f"""{model_prefix}encoder.layer.{layer_num}""" ) UpperCamelCase :Tuple = layer.transformer.layer[transformer_num].attention.attention.all_head_size UpperCamelCase :int = ( f"""{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.""" ) if "weight" in key: UpperCamelCase :Dict = val[:dim, :] UpperCamelCase :List[str] = val[dim : dim * 2, :] UpperCamelCase :Dict = val[-dim:, :] else: UpperCamelCase :Optional[int] = val[:dim] UpperCamelCase :List[Any] = val[dim : dim * 2] UpperCamelCase :Optional[int] = val[-dim:] else: UpperCamelCase :Union[str, Any] = val return orig_state_dict def SCREAMING_SNAKE_CASE_ ( ) -> List[str]: """simple docstring""" UpperCamelCase :Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCamelCase :Dict = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ ( __magic_name__ : List[str] , __magic_name__ : Optional[int] , __magic_name__ : List[str] , __magic_name__ : Optional[int]=False ) -> int: """simple docstring""" UpperCamelCase :Optional[int] = get_mobilevit_config(__magic_name__ ) # load original state_dict UpperCamelCase :Optional[Any] = torch.load(__magic_name__ , map_location="""cpu""" ) # load 🤗 model if mobilevit_name.startswith("""deeplabv3_""" ): UpperCamelCase :Union[str, Any] = MobileViTForSemanticSegmentation(__magic_name__ ).eval() else: UpperCamelCase :Optional[int] = MobileViTForImageClassification(__magic_name__ ).eval() UpperCamelCase :List[str] = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) # Check outputs on an image, prepared by MobileViTImageProcessor UpperCamelCase :Tuple = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) UpperCamelCase :Any = image_processor(images=prepare_img() , return_tensors="""pt""" ) UpperCamelCase :int = model(**__magic_name__ ) UpperCamelCase :Tuple = outputs.logits if mobilevit_name.startswith("""deeplabv3_""" ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": UpperCamelCase :Optional[int] = torch.tensor( [ [[6.2065, 6.1292, 6.2070], [6.1079, 6.1254, 6.1747], [6.0042, 6.1071, 6.1034]], [[-6.9253, -6.8653, -7.0398], [-7.3218, -7.3983, -7.3670], [-7.1961, -7.2482, -7.1569]], [[-4.4723, -4.4348, -4.3769], [-5.3629, -5.4632, -5.4598], [-5.1587, -5.3402, -5.5059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": UpperCamelCase :Tuple = torch.tensor( [ [[5.4449, 5.5733, 5.6314], [5.1815, 5.3930, 5.5963], [5.1656, 5.4333, 5.4853]], [[-9.4423, -9.7766, -9.6714], [-9.1581, -9.5720, -9.5519], [-9.1006, -9.6458, -9.5703]], [[-7.7721, -7.3716, -7.1583], [-8.4599, -8.0624, -7.7944], [-8.4172, -7.8366, -7.5025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": UpperCamelCase :Optional[Any] = torch.tensor( [ [[6.9811, 6.9743, 7.3123], [7.1777, 7.1931, 7.3938], [7.5633, 7.8050, 7.8901]], [[-10.5536, -10.2332, -10.2924], [-10.2336, -9.8624, -9.5964], [-10.8840, -10.8158, -10.6659]], [[-3.4938, -3.0631, -2.8620], [-3.4205, -2.8135, -2.6875], [-3.4179, -2.7945, -2.8750]], ] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3, :3, :3] , __magic_name__ , atol=1E-4 ) else: assert logits.shape == (1, 1000) if mobilevit_name == "mobilevit_s": UpperCamelCase :str = torch.tensor([-0.9866, 0.2392, -1.1241] ) elif mobilevit_name == "mobilevit_xs": UpperCamelCase :int = torch.tensor([-2.4761, -0.9399, -1.9587] ) elif mobilevit_name == "mobilevit_xxs": UpperCamelCase :Tuple = torch.tensor([-1.9364, -1.2327, -0.4653] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3] , __magic_name__ , atol=1E-4 ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) print(f"""Saving model {mobilevit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__magic_name__ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: UpperCamelCase :Dict = { """mobilevit_s""": """mobilevit-small""", """mobilevit_xs""": """mobilevit-x-small""", """mobilevit_xxs""": """mobilevit-xx-small""", """deeplabv3_mobilevit_s""": """deeplabv3-mobilevit-small""", """deeplabv3_mobilevit_xs""": """deeplabv3-mobilevit-x-small""", """deeplabv3_mobilevit_xxs""": """deeplabv3-mobilevit-xx-small""", } print("""Pushing to the hub...""" ) UpperCamelCase :Dict = model_mapping[mobilevit_name] image_processor.push_to_hub(__magic_name__ , organization="""apple""" ) model.push_to_hub(__magic_name__ , organization="""apple""" ) if __name__ == "__main__": UpperCAmelCase_ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--mobilevit_name''', default='''mobilevit_s''', type=str, help=( '''Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',''' ''' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.''' ), ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) UpperCAmelCase_ : int = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
38
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case : Dict =logging.get_logger(__name__) __snake_case : Optional[int] ={ 'facebook/xlm-roberta-xl': 'https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json', 'facebook/xlm-roberta-xxl': 'https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class lowerCamelCase__ ( lowerCamelCase__): '''simple docstring''' snake_case_ ="""xlm-roberta-xl""" def __init__(self ,__lowerCamelCase=25_08_80 ,__lowerCamelCase=25_60 ,__lowerCamelCase=36 ,__lowerCamelCase=32 ,__lowerCamelCase=1_02_40 ,__lowerCamelCase="gelu" ,__lowerCamelCase=0.1 ,__lowerCamelCase=0.1 ,__lowerCamelCase=5_14 ,__lowerCamelCase=1 ,__lowerCamelCase=0.02 ,__lowerCamelCase=1e-05 ,__lowerCamelCase=1 ,__lowerCamelCase=0 ,__lowerCamelCase=2 ,__lowerCamelCase="absolute" ,__lowerCamelCase=True ,__lowerCamelCase=None ,**__lowerCamelCase ,) -> Optional[Any]: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase ,bos_token_id=__lowerCamelCase ,eos_token_id=__lowerCamelCase ,**__lowerCamelCase ) lowerCAmelCase__ : int = vocab_size lowerCAmelCase__ : Tuple = hidden_size lowerCAmelCase__ : int = num_hidden_layers lowerCAmelCase__ : int = num_attention_heads lowerCAmelCase__ : int = hidden_act lowerCAmelCase__ : Tuple = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : Optional[Any] = max_position_embeddings lowerCAmelCase__ : List[str] = type_vocab_size lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Any = layer_norm_eps lowerCAmelCase__ : Union[str, Any] = position_embedding_type lowerCAmelCase__ : Union[str, Any] = use_cache lowerCAmelCase__ : str = classifier_dropout class lowerCamelCase__ ( lowerCamelCase__): '''simple docstring''' @property def lowerCAmelCase__ (self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": lowerCAmelCase__ : Dict = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase__ : Any = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
129
0
def UpperCAmelCase_ ( __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str] ) -> Dict: return int((input_a, input_a).count(0 ) != 0 ) def UpperCAmelCase_ ( ) -> Tuple: assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
355
from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def UpperCAmelCase_ ( __UpperCAmelCase : Optional[int] ) -> int: if not is_accelerate_available(): return method SCREAMING_SNAKE_CASE_ = version.parse(accelerate.__version__ ).base_version if version.parse(__UpperCAmelCase ) < version.parse('0.17.0' ): return method def wrapper(self : Optional[int] , *__UpperCAmelCase : Optional[Any] , **__UpperCAmelCase : Optional[Any] ): if hasattr(self , '_hf_hook' ) and hasattr(self._hf_hook , 'pre_forward' ): self._hf_hook.pre_forward(self ) return method(self , *__UpperCAmelCase , **__UpperCAmelCase ) return wrapper
210
0
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __magic_name__ ( datasets.BeamBasedBuilder): def UpperCAmelCase__ ( self : Tuple ) -> List[Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({'''content''': datasets.Value('''string''' )} ) , supervised_keys=lowerCamelCase__ , ) def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : str , lowerCamelCase__ : List[str] , lowerCamelCase__ : Dict ) -> str: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase__ ) class __magic_name__ ( datasets.BeamBasedBuilder): def UpperCAmelCase__ ( self : List[str] ) -> Any: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) , supervised_keys=lowerCamelCase__ , ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : Dict ) -> int: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : str , lowerCamelCase__ : Union[str, Any] ) -> List[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase__ ) def _a ( ): """simple docstring""" return [(i, {"content": content}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] def _a ( ): """simple docstring""" return [(i, {"a": {"b": [content]}}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] class __magic_name__ ( __lowerCAmelCase): @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[str] = DummyBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCamelCase__ : Any = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) self.assertDictEqual(dset['''train'''][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' import apache_beam as beam UpperCamelCase__ : List[Any] = beam.io.parquetio.WriteToParquet UpperCamelCase__ : str = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[Any] = DummyBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) with patch('''apache_beam.io.parquetio.WriteToParquet''' ) as write_parquet_mock: UpperCamelCase__ : Any = partial(lowerCamelCase__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCamelCase__ : Tuple = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset['''train''']['''content'''] ) , sorted(['''foo''', '''bar''', '''foobar'''] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Any ) -> Dict: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[Any] = DummyBeamDataset(cache_dir=lowerCamelCase__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : Tuple = NestedBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) ) UpperCamelCase__ : List[Any] = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) self.assertDictEqual(dset['''train'''][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset
146
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class __magic_name__ ( __lowerCAmelCase): def __init__( self : Optional[Any] , lowerCamelCase__ : Optional[NestedDataStructureLike[PathLike]] = None , lowerCamelCase__ : Optional[NamedSplit] = None , lowerCamelCase__ : Optional[Features] = None , lowerCamelCase__ : str = None , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : Optional[int] = None , **lowerCamelCase__ : int , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : int = path_or_paths UpperCamelCase__ : List[Any] = split if split or isinstance(lowerCamelCase__ , lowerCamelCase__ ) else '''train''' UpperCamelCase__ : Optional[Any] = features UpperCamelCase__ : List[Any] = cache_dir UpperCamelCase__ : Optional[int] = keep_in_memory UpperCamelCase__ : int = streaming UpperCamelCase__ : Union[str, Any] = num_proc UpperCamelCase__ : List[Any] = kwargs @abstractmethod def UpperCAmelCase__ ( self : Optional[Any] ) -> Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]: '''simple docstring''' pass class __magic_name__ ( __lowerCAmelCase): def __init__( self : int , lowerCamelCase__ : Optional[Features] = None , lowerCamelCase__ : str = None , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : Optional[int] = None , **lowerCamelCase__ : Union[str, Any] , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Optional[int] = features UpperCamelCase__ : Optional[int] = cache_dir UpperCamelCase__ : Union[str, Any] = keep_in_memory UpperCamelCase__ : Tuple = streaming UpperCamelCase__ : Optional[Any] = num_proc UpperCamelCase__ : Union[str, Any] = kwargs @abstractmethod def UpperCAmelCase__ ( self : Tuple ) -> Union[Dataset, IterableDataset]: '''simple docstring''' pass
146
1
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _UpperCamelCase = 50_0000 _UpperCamelCase , _UpperCamelCase = os.path.split(__file__) _UpperCamelCase = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def UpperCamelCase_( snake_case__: datasets.Dataset , **snake_case__: int ) -> Optional[Any]: UpperCAmelCase__ = dataset.map(**snake_case__ ) @get_duration def UpperCamelCase_( snake_case__: datasets.Dataset , **snake_case__: Any ) -> Union[str, Any]: UpperCAmelCase__ = dataset.filter(**snake_case__ ) def UpperCamelCase_( ) -> List[Any]: UpperCAmelCase__ = {'num examples': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = datasets.Features({'text': datasets.Value('string' ), 'numbers': datasets.Value('float32' )} ) UpperCAmelCase__ = generate_example_dataset( os.path.join(snake_case__ , 'dataset.arrow' ) , snake_case__ , num_examples=snake_case__ ) UpperCAmelCase__ = transformers.AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=snake_case__ ) def tokenize(snake_case__: Any ): return tokenizer(examples['text'] ) UpperCAmelCase__ = map(snake_case__ ) UpperCAmelCase__ = map(snake_case__ , batched=snake_case__ ) UpperCAmelCase__ = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type='numpy' ): UpperCAmelCase__ = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type='pandas' ): UpperCAmelCase__ = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type='torch' , columns='numbers' ): UpperCAmelCase__ = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type='tensorflow' , columns='numbers' ): UpperCAmelCase__ = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) UpperCAmelCase__ = map(snake_case__ , function=snake_case__ , batched=snake_case__ ) UpperCAmelCase__ = filter(snake_case__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(snake_case__ , 'wb' ) as f: f.write(json.dumps(snake_case__ ).encode('utf-8' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
335
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor _UpperCamelCase = logging.get_logger(__name__) class lowercase ( _UpperCamelCase ): '''simple docstring''' def __init__(self , *__a , **__a ) -> None: """simple docstring""" warnings.warn( 'The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use MobileViTImageProcessor instead.' , __a , ) super().__init__(*__a , **__a )
335
1
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __lowercase ( a_ ): """simple docstring""" UpperCamelCase : torch.FloatTensor class __lowercase ( nn.Module ): """simple docstring""" def __init__( self , A=3 , A=3 , A=("DownEncoderBlock2D",) , A=(64,) , A=2 , A=32 , A="silu" , A=True , ) -> Tuple: '''simple docstring''' super().__init__() lowerCamelCase = layers_per_block lowerCamelCase = torch.nn.Convad( A , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase = None lowerCamelCase = nn.ModuleList([] ) # down lowerCamelCase = block_out_channels[0] for i, down_block_type in enumerate(A ): lowerCamelCase = output_channel lowerCamelCase = block_out_channels[i] lowerCamelCase = i == len(A ) - 1 lowerCamelCase = get_down_block( A , num_layers=self.layers_per_block , in_channels=A , out_channels=A , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=A , resnet_groups=A , attention_head_dim=A , temb_channels=A , ) self.down_blocks.append(A ) # mid lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=A , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=A , temb_channels=A , ) # out lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=A , eps=1e-6 ) lowerCamelCase = nn.SiLU() lowerCamelCase = 2 * out_channels if double_z else out_channels lowerCamelCase = nn.Convad(block_out_channels[-1] , A , 3 , padding=1 ) lowerCamelCase = False def __A ( self , A ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = x lowerCamelCase = self.conv_in(A ) if self.training and self.gradient_checkpointing: def create_custom_forward(A ): def custom_forward(*A ): return module(*A ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(A ) , A , use_reentrant=A ) # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , A , use_reentrant=A ) else: for down_block in self.down_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(A ) , A ) # middle lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , A ) else: # down for down_block in self.down_blocks: lowerCamelCase = down_block(A ) # middle lowerCamelCase = self.mid_block(A ) # post-process lowerCamelCase = self.conv_norm_out(A ) lowerCamelCase = self.conv_act(A ) lowerCamelCase = self.conv_out(A ) return sample class __lowercase ( nn.Module ): """simple docstring""" def __init__( self , A=3 , A=3 , A=("UpDecoderBlock2D",) , A=(64,) , A=2 , A=32 , A="silu" , A="group" , ) -> Tuple: '''simple docstring''' super().__init__() lowerCamelCase = layers_per_block lowerCamelCase = nn.Convad( A , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase = None lowerCamelCase = nn.ModuleList([] ) lowerCamelCase = in_channels if norm_type == """spatial""" else None # mid lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=A , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=A , temb_channels=A , ) # up lowerCamelCase = list(reversed(A ) ) lowerCamelCase = reversed_block_out_channels[0] for i, up_block_type in enumerate(A ): lowerCamelCase = output_channel lowerCamelCase = reversed_block_out_channels[i] lowerCamelCase = i == len(A ) - 1 lowerCamelCase = get_up_block( A , num_layers=self.layers_per_block + 1 , in_channels=A , out_channels=A , prev_output_channel=A , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=A , resnet_groups=A , attention_head_dim=A , temb_channels=A , resnet_time_scale_shift=A , ) self.up_blocks.append(A ) lowerCamelCase = output_channel # out if norm_type == "spatial": lowerCamelCase = SpatialNorm(block_out_channels[0] , A ) else: lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=A , eps=1e-6 ) lowerCamelCase = nn.SiLU() lowerCamelCase = nn.Convad(block_out_channels[0] , A , 3 , padding=1 ) lowerCamelCase = False def __A ( self , A , A=None ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = z lowerCamelCase = self.conv_in(A ) lowerCamelCase = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(A ): def custom_forward(*A ): return module(*A ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , A , A , use_reentrant=A ) lowerCamelCase = sample.to(A ) # up for up_block in self.up_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(A ) , A , A , use_reentrant=A ) else: # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , A , A ) lowerCamelCase = sample.to(A ) # up for up_block in self.up_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(A ) , A , A ) else: # middle lowerCamelCase = self.mid_block(A , A ) lowerCamelCase = sample.to(A ) # up for up_block in self.up_blocks: lowerCamelCase = up_block(A , A ) # post-process if latent_embeds is None: lowerCamelCase = self.conv_norm_out(A ) else: lowerCamelCase = self.conv_norm_out(A , A ) lowerCamelCase = self.conv_act(A ) lowerCamelCase = self.conv_out(A ) return sample class __lowercase ( nn.Module ): """simple docstring""" def __init__( self , A , A , A , A=None , A="random" , A=False , A=True ) -> Any: '''simple docstring''' super().__init__() lowerCamelCase = n_e lowerCamelCase = vq_embed_dim lowerCamelCase = beta lowerCamelCase = legacy lowerCamelCase = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) lowerCamelCase = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) lowerCamelCase = self.used.shape[0] lowerCamelCase = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": lowerCamelCase = self.re_embed lowerCamelCase = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: lowerCamelCase = n_e lowerCamelCase = sane_index_shape def __A ( self , A ) -> Dict: '''simple docstring''' lowerCamelCase = inds.shape assert len(A ) > 1 lowerCamelCase = inds.reshape(ishape[0] , -1 ) lowerCamelCase = self.used.to(A ) lowerCamelCase = (inds[:, :, None] == used[None, None, ...]).long() lowerCamelCase = match.argmax(-1 ) lowerCamelCase = match.sum(2 ) < 1 if self.unknown_index == "random": lowerCamelCase = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: lowerCamelCase = self.unknown_index return new.reshape(A ) def __A ( self , A ) -> Any: '''simple docstring''' lowerCamelCase = inds.shape assert len(A ) > 1 lowerCamelCase = inds.reshape(ishape[0] , -1 ) lowerCamelCase = self.used.to(A ) if self.re_embed > self.used.shape[0]: # extra token lowerCamelCase = 0 # simply set to zero lowerCamelCase = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , A ) return back.reshape(A ) def __A ( self , A ) -> List[str]: '''simple docstring''' lowerCamelCase = z.permute(0 , 2 , 3 , 1 ).contiguous() lowerCamelCase = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z lowerCamelCase = torch.argmin(torch.cdist(A , self.embedding.weight ) , dim=1 ) lowerCamelCase = self.embedding(A ).view(z.shape ) lowerCamelCase = None lowerCamelCase = None # compute loss for embedding if not self.legacy: lowerCamelCase = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: lowerCamelCase = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients lowerCamelCase = z + (z_q - z).detach() # reshape back to match original input shape lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: lowerCamelCase = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis lowerCamelCase = self.remap_to_used(A ) lowerCamelCase = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: lowerCamelCase = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def __A ( self , A , A ) -> Dict: '''simple docstring''' if self.remap is not None: lowerCamelCase = indices.reshape(shape[0] , -1 ) # add batch axis lowerCamelCase = self.unmap_to_all(A ) lowerCamelCase = indices.reshape(-1 ) # flatten again # get quantized latent vectors lowerCamelCase = self.embedding(A ) if shape is not None: lowerCamelCase = z_q.view(A ) # reshape back to match original input shape lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __lowercase ( a_ ): """simple docstring""" def __init__( self , A , A=False ) -> List[str]: '''simple docstring''' lowerCamelCase = parameters lowerCamelCase , lowerCamelCase = torch.chunk(A , 2 , dim=1 ) lowerCamelCase = torch.clamp(self.logvar , -30.0 , 20.0 ) lowerCamelCase = deterministic lowerCamelCase = torch.exp(0.5 * self.logvar ) lowerCamelCase = torch.exp(self.logvar ) if self.deterministic: lowerCamelCase = lowerCamelCase = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def __A ( self , A = None ) -> torch.FloatTensor: '''simple docstring''' lowerCamelCase = randn_tensor( self.mean.shape , generator=A , device=self.parameters.device , dtype=self.parameters.dtype ) lowerCamelCase = self.mean + self.std * sample return x def __A ( self , A=None ) -> List[Any]: '''simple docstring''' if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def __A ( self , A , A=[1, 2, 3] ) -> Any: '''simple docstring''' if self.deterministic: return torch.Tensor([0.0] ) lowerCamelCase = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=A ) def __A ( self ) -> Tuple: '''simple docstring''' return self.mean
252
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class __lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self , A , A=7 , A=3 , A=30 , A=4_00 , A=True , A=None , A=True , A=[0.5, 0.5, 0.5] , A=[0.5, 0.5, 0.5] , A=True , A=1 / 2_55 , A=True , ) -> str: '''simple docstring''' lowerCamelCase = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 13_33} lowerCamelCase = parent lowerCamelCase = batch_size lowerCamelCase = num_channels lowerCamelCase = min_resolution lowerCamelCase = max_resolution lowerCamelCase = do_resize lowerCamelCase = size lowerCamelCase = do_normalize lowerCamelCase = image_mean lowerCamelCase = image_std lowerCamelCase = do_rescale lowerCamelCase = rescale_factor lowerCamelCase = do_pad def __A ( self ) -> List[Any]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __A ( self , A , A=False ) -> List[Any]: '''simple docstring''' if not batched: lowerCamelCase = image_inputs[0] if isinstance(A , Image.Image ): lowerCamelCase , lowerCamelCase = image.size else: lowerCamelCase , lowerCamelCase = image.shape[1], image.shape[2] if w < h: lowerCamelCase = int(self.size["""shortest_edge"""] * h / w ) lowerCamelCase = self.size["""shortest_edge"""] elif w > h: lowerCamelCase = self.size["""shortest_edge"""] lowerCamelCase = int(self.size["""shortest_edge"""] * w / h ) else: lowerCamelCase = self.size["""shortest_edge"""] lowerCamelCase = self.size["""shortest_edge"""] else: lowerCamelCase = [] for image in image_inputs: lowerCamelCase , lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase = max(A , key=lambda A : item[0] )[0] lowerCamelCase = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __lowercase ( a_ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Union[str, Any] = YolosImageProcessor if is_vision_available() else None def __A ( self ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase = YolosImageProcessingTester(self ) @property def __A ( self ) -> List[Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ) -> str: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , """image_mean""" ) ) self.assertTrue(hasattr(A , """image_std""" ) ) self.assertTrue(hasattr(A , """do_normalize""" ) ) self.assertTrue(hasattr(A , """do_resize""" ) ) self.assertTrue(hasattr(A , """size""" ) ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , A ) def __A ( self ) -> Union[str, Any]: '''simple docstring''' pass def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> Any: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) lowerCamelCase = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowerCamelCase = image_processing_a.pad(A , return_tensors="""pt""" ) lowerCamelCase = image_processing_a(A , return_tensors="""pt""" ) self.assertTrue( torch.allclose(encoded_images_with_method["""pixel_values"""] , encoded_images["""pixel_values"""] , atol=1e-4 ) ) @slow def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: lowerCamelCase = json.loads(f.read() ) lowerCamelCase = {"""image_id""": 3_97_69, """annotations""": target} # encode them lowerCamelCase = YolosImageProcessor.from_pretrained("""hustvl/yolos-small""" ) lowerCamelCase = image_processing(images=A , annotations=A , return_tensors="""pt""" ) # verify pixel values lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , A ) lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , A ) ) # verify boxes lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , A ) lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , A , atol=1e-3 ) ) # verify image_id lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , A ) ) # verify is_crowd lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , A ) ) # verify class_labels lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , A ) ) # verify orig_size lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , A ) ) # verify size lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , A ) ) @slow def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: lowerCamelCase = json.loads(f.read() ) lowerCamelCase = {"""file_name""": """000000039769.png""", """image_id""": 3_97_69, """segments_info""": target} lowerCamelCase = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them lowerCamelCase = YolosImageProcessor(format="""coco_panoptic""" ) lowerCamelCase = image_processing(images=A , annotations=A , masks_path=A , return_tensors="""pt""" ) # verify pixel values lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , A ) lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , A ) ) # verify boxes lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , A ) lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , A , atol=1e-3 ) ) # verify image_id lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , A ) ) # verify is_crowd lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , A ) ) # verify class_labels lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , A ) ) # verify masks lowerCamelCase = 82_28_73 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , A ) # verify orig_size lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , A ) ) # verify size lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , A ) )
252
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE : str = { "configuration_x_clip": [ "XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "XCLIPConfig", "XCLIPTextConfig", "XCLIPVisionConfig", ], "processing_x_clip": ["XCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = [ "XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "XCLIPModel", "XCLIPPreTrainedModel", "XCLIPTextModel", "XCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys SCREAMING_SNAKE_CASE : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
360
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) class UpperCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , **UpperCamelCase_ ): requires_backends(self , ['''bs4'''] ) super().__init__(**UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Optional[int] = [] lowercase_ :Union[str, Any] = [] lowercase_ :Union[str, Any] = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag lowercase_ :Any = parent.find_all(child.name , recursive=UpperCamelCase_ ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(UpperCamelCase_ ) else next(i for i, s in enumerate(UpperCamelCase_ , 1 ) if s is child ) ) lowercase_ :str = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Dict = BeautifulSoup(UpperCamelCase_ , '''html.parser''' ) lowercase_ :Union[str, Any] = [] lowercase_ :Union[str, Any] = [] lowercase_ :List[Any] = [] for element in html_code.descendants: if type(UpperCamelCase_ ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue lowercase_ :Dict = html.unescape(UpperCamelCase_ ).strip() if not text_in_this_tag: continue all_doc_strings.append(UpperCamelCase_ ) lowercase_ , lowercase_ :Tuple = self.xpath_soup(UpperCamelCase_ ) stringaxtag_seq.append(UpperCamelCase_ ) stringaxsubs_seq.append(UpperCamelCase_ ) if len(UpperCamelCase_ ) != len(UpperCamelCase_ ): raise ValueError('''Number of doc strings and xtags does not correspond''' ) if len(UpperCamelCase_ ) != len(UpperCamelCase_ ): raise ValueError('''Number of doc strings and xsubs does not correspond''' ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :Union[str, Any] = '''''' for tagname, subs in zip(UpperCamelCase_ , UpperCamelCase_ ): xpath += f"/{tagname}" if subs != 0: xpath += f"[{subs}]" return xpath def __call__( self , UpperCamelCase_ ): lowercase_ :Dict = False # Check that strings has a valid type if isinstance(UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :Optional[Any] = True elif isinstance(UpperCamelCase_ , (list, tuple) ): if len(UpperCamelCase_ ) == 0 or isinstance(html_strings[0] , UpperCamelCase_ ): lowercase_ :Tuple = True if not valid_strings: raise ValueError( '''HTML strings must of type `str`, `List[str]` (batch of examples), ''' f"but is of type {type(UpperCamelCase_ )}." ) lowercase_ :List[Any] = bool(isinstance(UpperCamelCase_ , (list, tuple) ) and (isinstance(html_strings[0] , UpperCamelCase_ )) ) if not is_batched: lowercase_ :Dict = [html_strings] # Get nodes + xpaths lowercase_ :List[Any] = [] lowercase_ :List[str] = [] for html_string in html_strings: lowercase_ , lowercase_ , lowercase_ :List[str] = self.get_three_from_single(UpperCamelCase_ ) nodes.append(UpperCamelCase_ ) lowercase_ :str = [] for node, tag_list, sub_list in zip(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :str = self.construct_xpath(UpperCamelCase_ , UpperCamelCase_ ) xpath_strings.append(UpperCamelCase_ ) xpaths.append(UpperCamelCase_ ) # return as Dict lowercase_ :int = {'''nodes''': nodes, '''xpaths''': xpaths} lowercase_ :Optional[int] = BatchFeature(data=UpperCamelCase_ , tensor_type=UpperCamelCase_ ) return encoded_inputs
252
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple = True , _lowerCAmelCase : Dict = None , _lowerCAmelCase : Dict = 3_2 , _lowerCAmelCase : Union[str, Any] = True , _lowerCAmelCase : str = 1 / 2_5_5 , _lowerCAmelCase : Union[str, Any] = True , _lowerCAmelCase : Dict = True , _lowerCAmelCase : Any = [0.48_145_466, 0.4_578_275, 0.40_821_073] , _lowerCAmelCase : str = [0.26_862_954, 0.26_130_258, 0.27_577_711] , _lowerCAmelCase : Optional[int] = True , _lowerCAmelCase : int=7 , _lowerCAmelCase : str=3_0 , _lowerCAmelCase : int=4_0_0 , _lowerCAmelCase : int=3 , ) -> Dict: """simple docstring""" snake_case_ = parent snake_case_ = do_resize snake_case_ = size if size is not None else {"shortest_edge": 2_8_8} snake_case_ = size_divisor snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = do_center_crop snake_case_ = image_mean snake_case_ = image_std snake_case_ = do_pad snake_case_ = batch_size snake_case_ = num_channels snake_case_ = min_resolution snake_case_ = max_resolution def lowerCAmelCase__ ( self : Dict ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def lowerCAmelCase__ ( self : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : str=False ) -> Optional[Any]: """simple docstring""" if not batched: snake_case_ = self.size["shortest_edge"] snake_case_ = image_inputs[0] if isinstance(_UpperCAmelCase , Image.Image ): snake_case_ = image.size else: snake_case_ = image.shape[1], image.shape[2] snake_case_ = size / min(_UpperCAmelCase , _UpperCAmelCase ) if h < w: snake_case_ = size, scale * w else: snake_case_ = scale * h, size snake_case_ = int((1_3_3_3 / 8_0_0) * size ) if max(_UpperCAmelCase , _UpperCAmelCase ) > max_size: snake_case_ = max_size / max(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ = newh * scale snake_case_ = neww * scale snake_case_ = int(newh + 0.5 ), int(neww + 0.5 ) snake_case_ = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: snake_case_ = [] for image in image_inputs: snake_case_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ = max(_UpperCAmelCase , key=lambda _lowerCAmelCase : item[0] )[0] snake_case_ = max(_UpperCAmelCase , key=lambda _lowerCAmelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = BridgeTowerImageProcessor if is_vision_available() else None def lowerCAmelCase__ ( self : Union[str, Any] ) -> Dict: """simple docstring""" snake_case_ = BridgeTowerImageProcessingTester(self ) @property def lowerCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCAmelCase , "image_mean" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "image_std" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "do_resize" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "size" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "size_divisor" ) ) def lowerCAmelCase__ ( self : List[str] ) -> List[str]: """simple docstring""" pass def lowerCAmelCase__ ( self : List[str] ) -> List[Any]: """simple docstring""" # Initialize image processor snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ = image_processing(_UpperCAmelCase , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase , batched=_UpperCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase__ ( self : int ) -> Tuple: """simple docstring""" # Initialize image processor snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , numpify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , np.ndarray ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ = image_processing(_UpperCAmelCase , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase , batched=_UpperCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase__ ( self : Union[str, Any] ) -> int: """simple docstring""" # Initialize image processor snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , torchify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , torch.Tensor ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ = image_processing(_UpperCAmelCase , return_tensors="pt" ).pixel_values snake_case_ = self.image_processor_tester.get_expected_values(_UpperCAmelCase , batched=_UpperCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
159
'''simple docstring''' from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A : '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase=1_3 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=9_9 , _UpperCAmelCase=3_2 , _UpperCAmelCase=2 , _UpperCAmelCase=4 , _UpperCAmelCase=3_7 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=5_1_2 , _UpperCAmelCase=1_6 , _UpperCAmelCase=2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=3 , _UpperCAmelCase=4 , _UpperCAmelCase=None , ) -> Dict: __UpperCamelCase : Optional[Any] = parent __UpperCamelCase : List[str] = 1_3 __UpperCamelCase : List[Any] = 7 __UpperCamelCase : List[str] = True __UpperCamelCase : Optional[Any] = True __UpperCamelCase : Tuple = True __UpperCamelCase : str = True __UpperCamelCase : List[Any] = 9_9 __UpperCamelCase : Union[str, Any] = 3_8_4 __UpperCamelCase : str = 2 __UpperCamelCase : Optional[Any] = 4 __UpperCamelCase : Any = 3_7 __UpperCamelCase : str = "gelu" __UpperCamelCase : Optional[Any] = 0.1 __UpperCamelCase : str = 0.1 __UpperCamelCase : str = 5_1_2 __UpperCamelCase : Optional[Any] = 1_6 __UpperCamelCase : Dict = 2 __UpperCamelCase : Optional[int] = 0.02 __UpperCamelCase : List[Any] = 3 __UpperCamelCase : Optional[Any] = 4 __UpperCamelCase : int = 1_2_8 __UpperCamelCase : Tuple = 2 __UpperCamelCase : str = 9 __UpperCamelCase : List[Any] = 1 __UpperCamelCase : Any = None def a_ (self ) -> int: __UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCamelCase : str = None if self.use_input_mask: __UpperCamelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase : int = None if self.use_token_type_ids: __UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCamelCase : List[Any] = None __UpperCamelCase : Union[str, Any] = None __UpperCamelCase : Optional[Any] = None if self.use_labels: __UpperCamelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCamelCase : Tuple = ids_tensor([self.batch_size] , self.num_choices ) __UpperCamelCase : str = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=_UpperCAmelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Dict: __UpperCamelCase : Tuple = TFConvBertModel(config=_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} __UpperCamelCase : Optional[Any] = [input_ids, input_mask] __UpperCamelCase : str = model(_UpperCAmelCase ) __UpperCamelCase : int = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: __UpperCamelCase : int = TFConvBertForMaskedLM(config=_UpperCAmelCase ) __UpperCamelCase : Dict = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __UpperCamelCase : List[str] = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: __UpperCamelCase : Union[str, Any] = self.num_labels __UpperCamelCase : Optional[Any] = TFConvBertForSequenceClassification(config=_UpperCAmelCase ) __UpperCamelCase : List[str] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __UpperCamelCase : Optional[Any] = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: __UpperCamelCase : Optional[int] = self.num_choices __UpperCamelCase : List[Any] = TFConvBertForMultipleChoice(config=_UpperCAmelCase ) __UpperCamelCase : Optional[int] = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __UpperCamelCase : Optional[Any] = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __UpperCamelCase : str = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } __UpperCamelCase : int = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: __UpperCamelCase : List[str] = self.num_labels __UpperCamelCase : Tuple = TFConvBertForTokenClassification(config=_UpperCAmelCase ) __UpperCamelCase : Dict = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __UpperCamelCase : Union[str, Any] = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: __UpperCamelCase : int = TFConvBertForQuestionAnswering(config=_UpperCAmelCase ) __UpperCamelCase : Dict = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __UpperCamelCase : Any = model(_UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a_ (self ) -> str: __UpperCamelCase : str = self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) : Any = config_and_inputs __UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' A = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) A = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) A = False A = False A = False def a_ (self ) -> Optional[int]: __UpperCamelCase : Tuple = TFConvBertModelTester(self ) __UpperCamelCase : Optional[Any] = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=3_7 ) def a_ (self ) -> Dict: self.config_tester.run_common_tests() def a_ (self ) -> Dict: __UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def a_ (self ) -> Tuple: __UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCAmelCase ) def a_ (self ) -> Tuple: __UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCAmelCase ) def a_ (self ) -> Dict: __UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCAmelCase ) def a_ (self ) -> Dict: __UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCAmelCase ) def a_ (self ) -> Optional[int]: __UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCAmelCase ) @slow def a_ (self ) -> Any: __UpperCamelCase , __UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase : str = True __UpperCamelCase : int = True if hasattr(_UpperCAmelCase , "use_cache" ): __UpperCamelCase : List[Any] = True __UpperCamelCase : List[str] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) __UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , _UpperCAmelCase ) for model_class in self.all_model_classes: __UpperCamelCase : Any = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase : int = model_class(_UpperCAmelCase ) __UpperCamelCase : Any = len(model(_UpperCAmelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_UpperCAmelCase , saved_model=_UpperCAmelCase ) __UpperCamelCase : List[str] = os.path.join(_UpperCAmelCase , "saved_model" , "1" ) __UpperCamelCase : List[str] = tf.keras.models.load_model(_UpperCAmelCase ) __UpperCamelCase : Dict = model(_UpperCAmelCase ) if self.is_encoder_decoder: __UpperCamelCase : Any = outputs["encoder_hidden_states"] __UpperCamelCase : Tuple = outputs["encoder_attentions"] else: __UpperCamelCase : Tuple = outputs["hidden_states"] __UpperCamelCase : Optional[int] = outputs["attentions"] self.assertEqual(len(_UpperCAmelCase ) , _UpperCAmelCase ) __UpperCamelCase : Any = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def a_ (self ) -> Optional[Any]: __UpperCamelCase : Tuple = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(_UpperCAmelCase ) def a_ (self ) -> Tuple: __UpperCamelCase , __UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase : str = True __UpperCamelCase : Tuple = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) __UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) __UpperCamelCase : Any = getattr(self.model_tester , "key_length" , _UpperCAmelCase ) __UpperCamelCase : List[Any] = getattr(self.model_tester , "key_length" , _UpperCAmelCase ) def check_decoder_attentions_output(_UpperCAmelCase ): __UpperCamelCase : Dict = len(_UpperCAmelCase ) self.assertEqual(out_len % 2 , 0 ) __UpperCamelCase : List[str] = outputs.decoder_attentions self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(_UpperCAmelCase ): __UpperCamelCase : Any = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: __UpperCamelCase : Any = True __UpperCamelCase : Dict = False __UpperCamelCase : str = model_class(_UpperCAmelCase ) __UpperCamelCase : Tuple = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) __UpperCamelCase : List[Any] = len(_UpperCAmelCase ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) if self.is_encoder_decoder: __UpperCamelCase : str = model_class(_UpperCAmelCase ) __UpperCamelCase : Dict = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_decoder_attentions_output(_UpperCAmelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] __UpperCamelCase : Optional[Any] = True __UpperCamelCase : Tuple = model_class(_UpperCAmelCase ) __UpperCamelCase : int = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) # Check attention is always last and order is fine __UpperCamelCase : int = True __UpperCamelCase : str = True __UpperCamelCase : Optional[Any] = model_class(_UpperCAmelCase ) __UpperCamelCase : Optional[int] = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_UpperCAmelCase ) ) self.assertEqual(model.config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) @require_tf class A ( unittest.TestCase ): '''simple docstring''' @slow def a_ (self ) -> str: __UpperCamelCase : Dict = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) __UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) __UpperCamelCase : Optional[int] = model(_UpperCAmelCase )[0] __UpperCamelCase : Tuple = [1, 6, 7_6_8] self.assertEqual(output.shape , _UpperCAmelCase ) __UpperCamelCase : Any = tf.constant( [ [ [-0.03_475_493, -0.4_686_034, -0.30_638_832], [0.22_637_248, -0.26_988_646, -0.7_423_424], [0.10_324_868, -0.45_013_508, -0.58_280_784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCAmelCase , atol=1E-4 )
298
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase__ = { "configuration_groupvit": [ "GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GroupViTConfig", "GroupViTOnnxConfig", "GroupViTTextConfig", "GroupViTVisionConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "GroupViTModel", "GroupViTPreTrainedModel", "GroupViTTextModel", "GroupViTVisionModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFGroupViTModel", "TFGroupViTPreTrainedModel", "TFGroupViTTextModel", "TFGroupViTVisionModel", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
360
'''simple docstring''' import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> str: '''simple docstring''' with open(SCREAMING_SNAKE_CASE__ ) as metadata_file: snake_case : int = json.load(SCREAMING_SNAKE_CASE__ ) snake_case : Any = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE__ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path snake_case : Any = torch.load(SCREAMING_SNAKE_CASE__ , map_location='''cpu''' )['''module'''] # Load the entity vocab file snake_case : Dict = load_original_entity_vocab(SCREAMING_SNAKE_CASE__ ) # add an entry for [MASK2] snake_case : List[str] = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 snake_case : int = XLMRobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks snake_case : Union[str, Any] = AddedToken('''<ent>''' , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) snake_case : Optional[int] = AddedToken('''<ent2>''' , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'Saving tokenizer to {pytorch_dump_folder_path}' ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , '''tokenizer_config.json''' ) , '''r''' ) as f: snake_case : Tuple = json.load(SCREAMING_SNAKE_CASE__ ) snake_case : List[str] = '''MLukeTokenizer''' with open(os.path.join(SCREAMING_SNAKE_CASE__ , '''tokenizer_config.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , MLukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) snake_case : List[Any] = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) # Initialize the embeddings of the special tokens snake_case : List[str] = tokenizer.convert_tokens_to_ids(['''@'''] )[0] snake_case : List[str] = tokenizer.convert_tokens_to_ids(['''#'''] )[0] snake_case : List[str] = state_dict['''embeddings.word_embeddings.weight'''] snake_case : int = word_emb[ent_init_index].unsqueeze(0 ) snake_case : Union[str, Any] = word_emb[enta_init_index].unsqueeze(0 ) snake_case : Dict = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: snake_case : Dict = state_dict[bias_name] snake_case : Any = decoder_bias[ent_init_index].unsqueeze(0 ) snake_case : str = decoder_bias[enta_init_index].unsqueeze(0 ) snake_case : Any = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: snake_case : Optional[Any] = F'encoder.layer.{layer_index}.attention.self.' snake_case : int = state_dict[prefix + matrix_name] snake_case : Union[str, Any] = state_dict[prefix + matrix_name] snake_case : int = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks snake_case : List[Any] = state_dict['''entity_embeddings.entity_embeddings.weight'''] snake_case : Dict = entity_emb[entity_vocab['''[MASK]''']].unsqueeze(0 ) snake_case : List[Any] = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' snake_case : Optional[Any] = state_dict['''entity_predictions.bias'''] snake_case : Optional[int] = entity_prediction_bias[entity_vocab['''[MASK]''']].unsqueeze(0 ) snake_case : List[Any] = torch.cat([entity_prediction_bias, entity_mask_bias] ) snake_case : str = LukeForMaskedLM(config=SCREAMING_SNAKE_CASE__ ).eval() state_dict.pop('''entity_predictions.decoder.weight''' ) state_dict.pop('''lm_head.decoder.weight''' ) state_dict.pop('''lm_head.decoder.bias''' ) snake_case : Optional[Any] = OrderedDict() for key, value in state_dict.items(): if not (key.startswith('''lm_head''' ) or key.startswith('''entity_predictions''' )): snake_case : int = state_dict[key] else: snake_case : List[str] = state_dict[key] snake_case ,snake_case : int = model.load_state_dict(SCREAMING_SNAKE_CASE__ , strict=SCREAMING_SNAKE_CASE__ ) if set(SCREAMING_SNAKE_CASE__ ) != {"luke.embeddings.position_ids"}: raise ValueError(F'Unexpected unexpected_keys: {unexpected_keys}' ) if set(SCREAMING_SNAKE_CASE__ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F'Unexpected missing_keys: {missing_keys}' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs snake_case : Optional[int] = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ , task='''entity_classification''' ) snake_case : Tuple = '''ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).''' snake_case : int = (0, 9) snake_case : str = tokenizer(SCREAMING_SNAKE_CASE__ , entity_spans=[span] , return_tensors='''pt''' ) snake_case : Union[str, Any] = model(**SCREAMING_SNAKE_CASE__ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base snake_case : Dict = torch.Size((1, 33, 768) ) snake_case : int = torch.tensor([[0.0892, 0.0596, -0.2819], [0.0134, 0.1199, 0.0573], [-0.0169, 0.0927, 0.0644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base snake_case : str = torch.Size((1, 1, 768) ) snake_case : Tuple = torch.tensor([[-0.1482, 0.0609, 0.0322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F'Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is' F' {expected_shape}' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ): raise ValueError # Verify masked word/entity prediction snake_case : str = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) snake_case : List[Any] = '''Tokyo is the capital of <mask>.''' snake_case : Union[str, Any] = (24, 30) snake_case : Tuple = tokenizer(SCREAMING_SNAKE_CASE__ , entity_spans=[span] , return_tensors='''pt''' ) snake_case : int = model(**SCREAMING_SNAKE_CASE__ ) snake_case : List[str] = encoding['''input_ids'''][0].tolist() snake_case : Union[str, Any] = input_ids.index(tokenizer.convert_tokens_to_ids('''<mask>''' ) ) snake_case : Dict = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(SCREAMING_SNAKE_CASE__ ) snake_case : List[Any] = outputs.entity_logits[0][0].argmax().item() snake_case : Dict = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith('''en:''' )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(SCREAMING_SNAKE_CASE__ ) ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> List[str]: '''simple docstring''' snake_case : Dict = ['''[MASK]''', '''[PAD]''', '''[UNK]'''] snake_case : List[Any] = [json.loads(SCREAMING_SNAKE_CASE__ ) for line in open(SCREAMING_SNAKE_CASE__ )] snake_case : Optional[int] = {} for entry in data: snake_case : Optional[Any] = entry['''id'''] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: snake_case : List[str] = entity_id break snake_case : Any = F'{language}:{entity_name}' snake_case : List[str] = entity_id return new_mapping if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument("--checkpoint_path", type=str, help="Path to a pytorch_model.bin file.") parser.add_argument( "--metadata_path", default=None, type=str, help="Path to a metadata.json file, defining the configuration." ) parser.add_argument( "--entity_vocab_path", default=None, type=str, help="Path to an entity_vocab.tsv file, containing the entity vocabulary.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to where to dump the output PyTorch model." ) parser.add_argument( "--model_size", default="base", type=str, choices=["base", "large"], help="Size of the model to be converted." ) lowercase__ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
83
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = { """configuration_autoformer""": [ """AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AutoformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ """AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """AutoformerForPrediction""", """AutoformerModel""", """AutoformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
323
'''simple docstring''' def __A ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" return number | (1 << position) def __A ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" return number & ~(1 << position) def __A ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" return number ^ (1 << position) def __A ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" return ((number >> position) & 1) == 1 def __A ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
323
1
"""simple docstring""" from __future__ import annotations from math import pi, sqrt def lowercase__ ( snake_case_ :float , snake_case_ :float ): if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
86
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging _lowercase : List[Any] = logging.get_logger(__name__) _lowercase : int = { 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/resolve/main/config.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/config.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/config.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json', } class _UpperCAmelCase ( _lowerCAmelCase ): a__ : Optional[Any] = "bloom" a__ : List[Any] = ["past_key_values"] a__ : Optional[Any] = { "num_hidden_layers": "n_layer", "num_attention_heads": "n_head", } def __init__( self : Union[str, Any] , _lowercase : Dict=25_08_80 , _lowercase : str=64 , _lowercase : int=2 , _lowercase : Union[str, Any]=8 , _lowercase : Optional[Any]=1E-5 , _lowercase : Dict=0.02 , _lowercase : Optional[int]=True , _lowercase : Any=1 , _lowercase : Dict=2 , _lowercase : Optional[Any]=False , _lowercase : Union[str, Any]=0.0 , _lowercase : str=0.0 , _lowercase : str=1 , _lowercase : int=False , **_lowercase : List[str] , ): __UpperCAmelCase = vocab_size # Backward compatibility with n_embed kwarg __UpperCAmelCase = kwargs.pop('''n_embed''' , _lowercase ) __UpperCAmelCase = hidden_size if n_embed is None else n_embed __UpperCAmelCase = n_layer __UpperCAmelCase = n_head __UpperCAmelCase = layer_norm_epsilon __UpperCAmelCase = initializer_range __UpperCAmelCase = use_cache __UpperCAmelCase = pretraining_tp __UpperCAmelCase = apply_residual_connection_post_layernorm __UpperCAmelCase = hidden_dropout __UpperCAmelCase = attention_dropout __UpperCAmelCase = bos_token_id __UpperCAmelCase = eos_token_id __UpperCAmelCase = slow_but_exact super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) class _UpperCAmelCase ( _lowerCAmelCase ): a__ : List[str] = version.parse("1.12" ) def __init__( self : Optional[int] , _lowercase : PretrainedConfig , _lowercase : str = "default" , _lowercase : List[PatchingSpec] = None , _lowercase : bool = False , ): super().__init__(_lowercase , task=_lowercase , patching_specs=_lowercase , use_past=_lowercase ) if not getattr(self._config , '''pad_token_id''' , _lowercase ): # TODO: how to do that better? __UpperCAmelCase = 0 @property def a ( self : Optional[int] ): __UpperCAmelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(_lowercase , direction='''inputs''' , inverted_values_shape=_lowercase ) __UpperCAmelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: __UpperCAmelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def a ( self : Any ): return self._config.n_layer @property def a ( self : Tuple ): return self._config.n_head @property def a ( self : Dict ): return 1E-3 def a ( self : List[str] , _lowercase : "PreTrainedTokenizer" , _lowercase : int = -1 , _lowercase : int = -1 , _lowercase : bool = False , _lowercase : Optional["TensorType"] = None , ): __UpperCAmelCase = super(_lowercase , self ).generate_dummy_inputs( _lowercase , batch_size=_lowercase , seq_length=_lowercase , is_pair=_lowercase , framework=_lowercase ) # We need to order the input in the way they appears in the forward() __UpperCAmelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch __UpperCAmelCase , __UpperCAmelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values __UpperCAmelCase = seqlen + 2 __UpperCAmelCase = self._config.hidden_size // self.num_attention_heads __UpperCAmelCase = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) __UpperCAmelCase = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) __UpperCAmelCase = [ (torch.zeros(_lowercase ), torch.zeros(_lowercase )) for _ in range(self.num_layers ) ] __UpperCAmelCase = common_inputs['''attention_mask'''] if self.use_past: __UpperCAmelCase = ordered_inputs['''attention_mask'''].dtype __UpperCAmelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(_lowercase , _lowercase , dtype=_lowercase )] , dim=1 ) return ordered_inputs @property def a ( self : Any ): return 13
86
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING import torch from ..models.auto import AutoModelForVisualQuestionAnswering, AutoProcessor from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = """dandelin/vilt-b32-finetuned-vqa""" SCREAMING_SNAKE_CASE_ : Tuple = ( """This is a tool that answers a question about an image. It takes an input named `image` which should be the """ """image containing the information, as well as a `question` which should be the question in English. It """ """returns a text that is the answer to the question.""" ) SCREAMING_SNAKE_CASE_ : Tuple = """image_qa""" SCREAMING_SNAKE_CASE_ : List[str] = AutoProcessor SCREAMING_SNAKE_CASE_ : Any = AutoModelForVisualQuestionAnswering SCREAMING_SNAKE_CASE_ : str = ["""image""", """text"""] SCREAMING_SNAKE_CASE_ : Union[str, Any] = ["""text"""] def __init__( self : int , *__lowerCamelCase : int , **__lowerCamelCase : Dict ) -> str: requires_backends(self , ["vision"] ) super().__init__(*__lowerCamelCase , **__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : "Image" , __lowerCamelCase : str ) -> Union[str, Any]: return self.pre_processor(__lowerCamelCase , __lowerCamelCase , return_tensors="pt" ) def __UpperCAmelCase ( self : int , __lowerCamelCase : Dict ) -> List[str]: with torch.no_grad(): return self.model(**__lowerCamelCase ).logits def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : List[Any] ) -> Any: a = outputs.argmax(-1 ).item() return self.model.config.idalabel[idx]
107
"""simple docstring""" from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function __UpperCAmelCase = 1.054571817e-34 # unit of ℏ : J * s __UpperCAmelCase = 3e8 # unit of c : m * s^-1 def _snake_case ( lowercase__ : float , lowercase__ : float , lowercase__ : float ) -> dict[str, float]: '''simple docstring''' if (force, area, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if force < 0: raise ValueError("""Magnitude of force can not be negative""" ) if distance < 0: raise ValueError("""Distance can not be negative""" ) if area < 0: raise ValueError("""Area can not be negative""" ) if force == 0: lowerCAmelCase_ :Union[str, Any] = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 2_4_0 * (distance) ** 4 ) return {"force": force} elif area == 0: lowerCAmelCase_ :Optional[Any] = (2_4_0 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: lowerCAmelCase_ :Any = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (2_4_0 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError("""One and only one argument must be 0""" ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
84
0
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowercase_ = logging.get_logger(__name__) lowercase_ = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} lowercase_ = { "vocab_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/vocab.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/vocab.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/vocab.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/vocab.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/vocab.json", }, "merges_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/merges.txt", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/merges.txt", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/merges.txt", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/merges.txt", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/merges.txt", }, "tokenizer_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/tokenizer.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/tokenizer.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/tokenizer.json", }, } lowercase_ = { "gpt2": 1_0_2_4, "gpt2-medium": 1_0_2_4, "gpt2-large": 1_0_2_4, "gpt2-xl": 1_0_2_4, "distilgpt2": 1_0_2_4, } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : Optional[int] = VOCAB_FILES_NAMES __UpperCAmelCase : Any = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Union[str, Any] = ['input_ids', 'attention_mask'] __UpperCAmelCase : Optional[int] = GPTaTokenizer def __init__( self , _a=None , _a=None , _a=None , _a="<|endoftext|>" , _a="<|endoftext|>" , _a="<|endoftext|>" , _a=False , **_a , ): super().__init__( _a , _a , tokenizer_file=_a , unk_token=_a , bos_token=_a , eos_token=_a , add_prefix_space=_a , **_a , ) __a = kwargs.pop('''add_bos_token''' , _a ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , _a ) != add_prefix_space: __a = getattr(_a , pre_tok_state.pop('''type''' ) ) __a = add_prefix_space __a = pre_tok_class(**_a ) __a = add_prefix_space def __UpperCAmelCase ( self , *_a , **_a ): __a = kwargs.get('''is_split_into_words''' , _a ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*_a , **_a ) def __UpperCAmelCase ( self , *_a , **_a ): __a = kwargs.get('''is_split_into_words''' , _a ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*_a , **_a ) def __UpperCAmelCase ( self , _a , _a = None ): __a = self._tokenizer.model.save(_a , name=_a ) return tuple(_a ) def __UpperCAmelCase ( self , _a ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_a , add_special_tokens=_a ) + [self.eos_token_id] ) if len(_a ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
11
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging lowercase_ = logging.get_logger(__name__) # pylint: disable=invalid-name class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , _a , _a , _a , _a , _a , _a , _a , _a , _a , ): super().__init__() if hasattr(scheduler.config , '''steps_offset''' ) and scheduler.config.steps_offset != 1: __a = ( f'''The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`''' f''' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ''' '''to update the config accordingly as leaving `steps_offset` might led to incorrect results''' ''' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,''' ''' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`''' ''' file''' ) deprecate('''steps_offset!=1''' , '''1.0.0''' , _a , standard_warn=_a ) __a = dict(scheduler.config ) __a = 1 __a = FrozenDict(_a ) if hasattr(scheduler.config , '''skip_prk_steps''' ) and scheduler.config.skip_prk_steps is False: __a = ( f'''The configuration file of this scheduler: {scheduler} has not set the configuration''' ''' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make''' ''' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to''' ''' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face''' ''' Hub, it would be very nice if you could open a Pull request for the''' ''' `scheduler/scheduler_config.json` file''' ) deprecate('''skip_prk_steps not set''' , '''1.0.0''' , _a , standard_warn=_a ) __a = dict(scheduler.config ) __a = True __a = FrozenDict(_a ) if safety_checker is None: logger.warning( f'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ''' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered''' ''' results in services or applications open to the public. Both the diffusers team and Hugging Face''' ''' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling''' ''' it only for use-cases that involve analyzing network behavior or auditing its results. For more''' ''' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .''' ) self.register_modules( segmentation_model=_a , segmentation_processor=_a , vae=_a , text_encoder=_a , tokenizer=_a , unet=_a , scheduler=_a , safety_checker=_a , feature_extractor=_a , ) def __UpperCAmelCase ( self , _a = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __a = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_a ) def __UpperCAmelCase ( self ): self.enable_attention_slicing(_a ) def __UpperCAmelCase ( self ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) __a = torch.device('''cuda''' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(_a , _a ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __UpperCAmelCase ( self ): if self.device != torch.device('''meta''' ) or not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(_a , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , _a , _a , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): __a = self.segmentation_processor( text=[text] , images=[image] , padding='''max_length''' , return_tensors='''pt''' ).to(self.device ) __a = self.segmentation_model(**_a ) __a = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __a = self.numpy_to_pil(_a )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __a = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=_a , image=_a , mask_image=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , )
11
1
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class _lowerCamelCase: lowercase_ : Optional[Union[str, Path]] = None lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = False lowercase_ : Optional[Dict] = None lowercase_ : Optional[str] = None lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = True lowercase_ : Optional[int] = None lowercase_ : int = 1 lowercase_ : Optional[Union[str, bool]] = None lowercase_ : bool = False lowercase_ : Optional[Dict] = None lowercase_ : Optional[str] = None def UpperCamelCase ( self) -> "DownloadConfig": """simple docstring""" return self.__class__(**{k: copy.deepcopy(lowerCamelCase) for k, v in self.__dict__.items()})
21
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def UpperCamelCase_( lowerCamelCase_ ) -> bool: _lowercase : int = int(number**0.5 ) return number == sq * sq def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> tuple[int, int]: _lowercase : int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den _lowercase : int = x_den * y_den * z_den _lowercase : int = gcd(lowerCamelCase_ , lowerCamelCase_ ) top //= hcf bottom //= hcf return top, bottom def UpperCamelCase_( lowerCamelCase_ = 35 ) -> int: _lowercase : set = set() _lowercase : int _lowercase : Fraction = Fraction(0 ) _lowercase : tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 _lowercase : int = x_num * y_den + x_den * y_num _lowercase : int = x_den * y_den _lowercase : str = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : List[Any] = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 _lowercase : Dict = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) _lowercase : List[Any] = x_den * x_den * y_den * y_den if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): _lowercase : Tuple = int(sqrt(lowerCamelCase_ ) ) _lowercase : int = int(sqrt(lowerCamelCase_ ) ) _lowercase : Any = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : Optional[int] = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=-1 _lowercase : Any = x_num * y_num _lowercase : str = x_den * y_num + x_num * y_den _lowercase : Any = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : int = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 _lowercase : str = x_num * x_num * y_num * y_num _lowercase : Optional[Any] = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): _lowercase : Tuple = int(sqrt(lowerCamelCase_ ) ) _lowercase : List[str] = int(sqrt(lowerCamelCase_ ) ) _lowercase : Union[str, Any] = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : Tuple = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) for num, den in unique_s: total += Fraction(lowerCamelCase_ , lowerCamelCase_ ) return total.denominator + total.numerator if __name__ == "__main__": print(F"{solution() = }")
21
1
"""simple docstring""" def a__ ( snake_case__ , snake_case__ ) -> int: return int(input_a == input_a == 0 ) def a__ ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(F'| 0 | 0 | {nor_gate(0 , 0 )} |' ) print(F'| 0 | 1 | {nor_gate(0 , 1 )} |' ) print(F'| 1 | 0 | {nor_gate(1 , 0 )} |' ) print(F'| 1 | 1 | {nor_gate(1 , 1 )} |' ) if __name__ == "__main__": import doctest doctest.testmod() main()
168
"""simple docstring""" def a__ ( snake_case__ , snake_case__ ) -> int: return number | (1 << position) def a__ ( snake_case__ , snake_case__ ) -> int: return number & ~(1 << position) def a__ ( snake_case__ , snake_case__ ) -> int: return number ^ (1 << position) def a__ ( snake_case__ , snake_case__ ) -> bool: return ((number >> position) & 1) == 1 def a__ ( snake_case__ , snake_case__ ) -> int: return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
168
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowercase : int = logging.get_logger(__name__) __lowercase : Optional[Any] = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __UpperCamelCase ( lowerCAmelCase_ ): A_ = "roberta-prelayernorm" def __init__( self , __a=5_0265 , __a=768 , __a=12 , __a=12 , __a=3072 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=2 , __a=0.02 , __a=1E-1_2 , __a=1 , __a=0 , __a=2 , __a="absolute" , __a=True , __a=None , **__a , ): '''simple docstring''' super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) __a : Optional[Any] = vocab_size __a : str = hidden_size __a : int = num_hidden_layers __a : Union[str, Any] = num_attention_heads __a : Any = hidden_act __a : Union[str, Any] = intermediate_size __a : int = hidden_dropout_prob __a : Optional[int] = attention_probs_dropout_prob __a : Any = max_position_embeddings __a : str = type_vocab_size __a : Tuple = initializer_range __a : Any = layer_norm_eps __a : List[str] = position_embedding_type __a : str = use_cache __a : str = classifier_dropout class __UpperCamelCase ( lowerCAmelCase_ ): @property def __UpperCAmelCase ( self ): '''simple docstring''' if self.task == "multiple-choice": __a : List[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __a : Any = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
27
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase = 100 ,) -> float: __lowerCamelCase : Dict = x_start __lowerCamelCase : int = fnc(_lowerCAmelCase ) __lowerCamelCase : Dict = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length __lowerCamelCase : List[str] = (x_end - x_start) / steps + xa __lowerCamelCase : List[Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa ,fxa - fxa ) # Increment step __lowerCamelCase : Any = xa __lowerCamelCase : Tuple = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase ) -> Dict: return math.sin(10 * x ) print('f(x) = sin(10 * x)') print('The length of the curve from x = -10 to x = 10 is:') _UpperCamelCase = 10 while i <= 100000: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
208
0
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def a__ ( UpperCAmelCase : Union[str, Any] ) -> List[Any]: return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def a__ ( UpperCAmelCase : Any ) -> int: UpperCAmelCase : int = create_tensor(UpperCAmelCase ) UpperCAmelCase : Union[str, Any] = gather(UpperCAmelCase ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def a__ ( UpperCAmelCase : int ) -> Tuple: UpperCAmelCase : Optional[int] = [state.process_index] UpperCAmelCase : Union[str, Any] = gather_object(UpperCAmelCase ) assert len(UpperCAmelCase ) == state.num_processes, f'''{gathered_obj}, {len(UpperCAmelCase )} != {state.num_processes}''' assert gathered_obj == list(range(state.num_processes ) ), f'''{gathered_obj} != {list(range(state.num_processes ) )}''' def a__ ( UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Tuple = create_tensor(UpperCAmelCase ) UpperCAmelCase : List[str] = broadcast(UpperCAmelCase ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def a__ ( UpperCAmelCase : Any ) -> Tuple: # We need to pad the tensor with one more element if we are the main process # to ensure that we can pad if state.is_main_process: UpperCAmelCase : Any = torch.arange(state.num_processes + 1 ).to(state.device ) else: UpperCAmelCase : int = torch.arange(state.num_processes ).to(state.device ) UpperCAmelCase : List[str] = pad_across_processes(UpperCAmelCase ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def a__ ( UpperCAmelCase : int ) -> List[str]: # For now runs on only two processes if state.num_processes != 2: return UpperCAmelCase : Tuple = create_tensor(UpperCAmelCase ) UpperCAmelCase : str = reduce(UpperCAmelCase , '''sum''' ) UpperCAmelCase : List[Any] = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(UpperCAmelCase , UpperCAmelCase ), f'''{reduced_tensor} != {truth_tensor}''' def a__ ( UpperCAmelCase : Tuple ) -> int: # For now runs on only two processes if state.num_processes != 2: return UpperCAmelCase : Dict = create_tensor(UpperCAmelCase ) UpperCAmelCase : List[str] = reduce(UpperCAmelCase , '''mean''' ) UpperCAmelCase : Union[str, Any] = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(UpperCAmelCase , UpperCAmelCase ), f'''{reduced_tensor} != {truth_tensor}''' def a__ ( UpperCAmelCase : Optional[Any] ) -> List[str]: # For xla_spawn (TPUs) main() def a__ ( ) -> Union[str, Any]: UpperCAmelCase : List[str] = PartialState() state.print(f'''State: {state}''' ) state.print('''testing gather''' ) test_gather(UpperCAmelCase ) state.print('''testing gather_object''' ) test_gather_object(UpperCAmelCase ) state.print('''testing broadcast''' ) test_broadcast(UpperCAmelCase ) state.print('''testing pad_across_processes''' ) test_pad_across_processes(UpperCAmelCase ) state.print('''testing reduce_sum''' ) test_reduce_sum(UpperCAmelCase ) state.print('''testing reduce_mean''' ) test_reduce_mean(UpperCAmelCase ) if __name__ == "__main__": main()
99
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase : str = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : List[Any] = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Union[str, Any] = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys _lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
99
1
import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() snake_case_ = logging.get_logger(__name__) snake_case_ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'encoder.layer_norm_for_extract': 'layer_norm_for_extract', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'label_embs_concat': 'label_embeddings_concat', 'mask_emb': 'masked_spec_embed', 'spk_proj': 'speaker_proj', } snake_case_ = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', 'label_embeddings_concat', 'speaker_proj', 'layer_norm_for_extract', ] def lowerCamelCase__ ( snake_case_ : Tuple , snake_case_ : Optional[Any] , snake_case_ : List[Any] , snake_case_ : List[str] , snake_case_ : str ) -> List[str]: for attribute in key.split('''.''' ): __snake_case = getattr(snake_case_ , snake_case_ ) if weight_type is not None: __snake_case = getattr(snake_case_ , snake_case_ ).shape else: __snake_case = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __snake_case = value elif weight_type == "weight_g": __snake_case = value elif weight_type == "weight_v": __snake_case = value elif weight_type == "bias": __snake_case = value else: __snake_case = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase__ ( snake_case_ : Any , snake_case_ : List[str] ) -> List[str]: __snake_case = [] __snake_case = fairseq_model.state_dict() __snake_case = hf_model.unispeech_sat.feature_extractor for name, value in fairseq_dict.items(): __snake_case = False if "conv_layers" in name: load_conv_layer( snake_case_ , snake_case_ , snake_case_ , snake_case_ , hf_model.config.feat_extract_norm == '''group''' , ) __snake_case = True else: for key, mapped_key in MAPPING.items(): __snake_case = '''unispeech_sat.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: if "layer_norm_for_extract" in name and (".".join(name.split('''.''' )[:-1] ) != key): # special case since naming is very similar continue __snake_case = True if "*" in mapped_key: __snake_case = name.split(snake_case_ )[0].split('''.''' )[-2] __snake_case = mapped_key.replace('''*''' , snake_case_ ) if "weight_g" in name: __snake_case = '''weight_g''' elif "weight_v" in name: __snake_case = '''weight_v''' elif "bias" in name: __snake_case = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __snake_case = '''weight''' else: __snake_case = None set_recursively(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) continue if not is_used: unused_weights.append(snake_case_ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase__ ( snake_case_ : str , snake_case_ : Union[str, Any] , snake_case_ : Tuple , snake_case_ : Tuple , snake_case_ : Optional[int] ) -> List[Any]: __snake_case = full_name.split('''conv_layers.''' )[-1] __snake_case = name.split('''.''' ) __snake_case = int(items[0] ) __snake_case = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __snake_case = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __snake_case = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.bias.data.shape} was found.""" ) __snake_case = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __snake_case = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(snake_case_ ) @torch.no_grad() def lowerCamelCase__ ( snake_case_ : Optional[Any] , snake_case_ : Union[str, Any] , snake_case_ : int=None , snake_case_ : str=None , snake_case_ : Optional[int]=True ) -> int: if config_path is not None: __snake_case = UniSpeechSatConfig.from_pretrained(snake_case_ ) else: __snake_case = UniSpeechSatConfig() __snake_case = '''''' if is_finetuned: __snake_case = UniSpeechSatForCTC(snake_case_ ) else: __snake_case = UniSpeechSatForPreTraining(snake_case_ ) __snake_case , __snake_case , __snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) __snake_case = model[0].eval() recursively_load_weights(snake_case_ , snake_case_ ) hf_wavavec.save_pretrained(snake_case_ ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case_ = parser.parse_args() convert_unispeech_sat_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
24
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: int ) -> list: '''simple docstring''' A__ = int(SCREAMING_SNAKE_CASE_ ) if n_element < 1: A__ = ValueError("a should be a positive number" ) raise my_error A__ = [1] A__ , A__ , A__ = (0, 0, 0) A__ = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": lowerCAmelCase__ = input("""Enter the last number (nth term) of the Hamming Number Series: """) print("""Formula of Hamming Number Series => 2^i * 3^j * 5^k""") lowerCAmelCase__ = hamming(int(n)) print("""-----------------------------------------------------""") print(f"""The list with nth numbers is: {hamming_numbers}""") print("""-----------------------------------------------------""")
68
0
def UpperCamelCase ( ) -> list[list[int]]: '''simple docstring''' return [list(range(1_0_0_0 - i , -1_0_0_0 - i , -1 ) ) for i in range(1_0_0_0 )] SCREAMING_SNAKE_CASE : Union[str, Any] = generate_large_matrix() SCREAMING_SNAKE_CASE : Tuple = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def UpperCamelCase ( _a ) -> None: '''simple docstring''' assert all(row == sorted(_a , reverse=_a ) for row in grid ) assert all(list(_a ) == sorted(_a , reverse=_a ) for col in zip(*_a ) ) def UpperCamelCase ( _a ) -> int: '''simple docstring''' lowercase_ :int = 0 lowercase_ :List[str] = len(_a ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: lowercase_ :Union[str, Any] = (left + right) // 2 lowercase_ :Optional[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: lowercase_ :Dict = mid + 1 else: lowercase_ :Optional[Any] = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(_a ) def UpperCamelCase ( _a ) -> int: '''simple docstring''' lowercase_ :int = 0 lowercase_ :Tuple = len(grid[0] ) for i in range(len(_a ) ): lowercase_ :Optional[Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(_a ) * len(grid[0] )) - total def UpperCamelCase ( _a ) -> int: '''simple docstring''' return len([number for row in grid for number in row if number < 0] ) def UpperCamelCase ( _a ) -> int: '''simple docstring''' lowercase_ :List[str] = 0 for row in grid: for i, number in enumerate(_a ): if number < 0: total += len(_a ) - i break return total def UpperCamelCase ( ) -> None: '''simple docstring''' from timeit import timeit print('''Running benchmarks''' ) lowercase_ :Union[str, Any] = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): lowercase_ :List[Any] = timeit(f"{func}(grid=grid)" , setup=_a , number=5_0_0 ) print(f"{func}() took {time:0.4f} seconds" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
252
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers SCREAMING_SNAKE_CASE : Dict = [ "python", "tqdm", "regex", "requests", "packaging", "filelock", "numpy", "tokenizers", "huggingface-hub", "safetensors", "accelerate", "pyyaml", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"can't find {pkg} in {deps.keys()}, check dependency_versions_table.py") def UpperCamelCase ( _a , _a=None ) -> Optional[int]: '''simple docstring''' require_version(deps[pkg] , _a )
252
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ : str = { """configuration_conditional_detr""": [ """CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConditionalDetrConfig""", """ConditionalDetrOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : str = ["""ConditionalDetrFeatureExtractor"""] a_ : Any = ["""ConditionalDetrImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Union[str, Any] = [ """CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConditionalDetrForObjectDetection""", """ConditionalDetrForSegmentation""", """ConditionalDetrModel""", """ConditionalDetrPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys a_ : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
55
"""simple docstring""" def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: """simple docstring""" if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) lowerCAmelCase__ :Union[str, Any] = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
293
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} UpperCAmelCase = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } UpperCAmelCase = { """google/realm-cc-news-pretrained-embedder""": 512, """google/realm-cc-news-pretrained-encoder""": 512, """google/realm-cc-news-pretrained-scorer""": 512, """google/realm-cc-news-pretrained-openqa""": 512, """google/realm-orqa-nq-openqa""": 512, """google/realm-orqa-nq-reader""": 512, """google/realm-orqa-wq-openqa""": 512, """google/realm-orqa-wq-reader""": 512, } UpperCAmelCase = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_INIT_CONFIGURATION __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = RealmTokenizer def __init__( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=True , _UpperCAmelCase="[UNK]" , _UpperCAmelCase="[SEP]" , _UpperCAmelCase="[PAD]" , _UpperCAmelCase="[CLS]" , _UpperCAmelCase="[MASK]" , _UpperCAmelCase=True , _UpperCAmelCase=None , **_UpperCAmelCase , ): super().__init__( _UpperCAmelCase , tokenizer_file=_UpperCAmelCase , do_lower_case=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , tokenize_chinese_chars=_UpperCAmelCase , strip_accents=_UpperCAmelCase , **_UpperCAmelCase , ) snake_case_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _UpperCAmelCase ) != do_lower_case or normalizer_state.get('''strip_accents''' , _UpperCAmelCase ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _UpperCAmelCase ) != tokenize_chinese_chars ): snake_case_ = getattr(_UpperCAmelCase , normalizer_state.pop('''type''' ) ) snake_case_ = do_lower_case snake_case_ = strip_accents snake_case_ = tokenize_chinese_chars snake_case_ = normalizer_class(**_UpperCAmelCase ) snake_case_ = do_lower_case def UpperCamelCase__ ( self , _UpperCAmelCase , **_UpperCAmelCase ): snake_case_ = PaddingStrategy.MAX_LENGTH snake_case_ = text snake_case_ = kwargs.pop('''text_pair''' , _UpperCAmelCase ) snake_case_ = kwargs.pop('''return_tensors''' , _UpperCAmelCase ) snake_case_ = { '''input_ids''': [], '''attention_mask''': [], '''token_type_ids''': [], } for idx, candidate_text in enumerate(_UpperCAmelCase ): if batch_text_pair is not None: snake_case_ = batch_text_pair[idx] else: snake_case_ = None snake_case_ = super().__call__(_UpperCAmelCase , _UpperCAmelCase , return_tensors=_UpperCAmelCase , **_UpperCAmelCase ) snake_case_ = encoded_candidates.get('''input_ids''' ) snake_case_ = encoded_candidates.get('''attention_mask''' ) snake_case_ = encoded_candidates.get('''token_type_ids''' ) if encoded_input_ids is not None: output_data["input_ids"].append(_UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_UpperCAmelCase ) snake_case_ = {key: item for key, item in output_data.items() if len(_UpperCAmelCase ) != 0} return BatchEncoding(_UpperCAmelCase , tensor_type=_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase=None ): snake_case_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None ): snake_case_ = self._tokenizer.model.save(_UpperCAmelCase , name=_UpperCAmelCase ) return tuple(_UpperCAmelCase )
267
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu UpperCAmelCase = False class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase__ ( self ): return 12 @property def UpperCamelCase__ ( self ): return 12 @property def UpperCamelCase__ ( self ): return 32 @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=3 , num_vq_embeddings=self.num_embed , vq_embed_dim=3 , ) return model @property def UpperCamelCase__ ( self ): snake_case_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(_UpperCAmelCase ) @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) snake_case_ = 12 snake_case_ = 12 snake_case_ = { '''attention_bias''': True, '''cross_attention_dim''': 32, '''attention_head_dim''': height * width, '''num_attention_heads''': 1, '''num_vector_embeds''': self.num_embed, '''num_embeds_ada_norm''': self.num_embeds_ada_norm, '''norm_num_groups''': 32, '''sample_size''': width, '''activation_fn''': '''geglu-approximate''', } snake_case_ = TransformeraDModel(**_UpperCAmelCase ) return model def UpperCamelCase__ ( self ): snake_case_ = '''cpu''' snake_case_ = self.dummy_vqvae snake_case_ = self.dummy_text_encoder snake_case_ = self.dummy_tokenizer snake_case_ = self.dummy_transformer snake_case_ = VQDiffusionScheduler(self.num_embed ) snake_case_ = LearnedClassifierFreeSamplingEmbeddings(learnable=_UpperCAmelCase ) snake_case_ = VQDiffusionPipeline( vqvae=_UpperCAmelCase , text_encoder=_UpperCAmelCase , tokenizer=_UpperCAmelCase , transformer=_UpperCAmelCase , scheduler=_UpperCAmelCase , learned_classifier_free_sampling_embeddings=_UpperCAmelCase , ) snake_case_ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) snake_case_ = '''teddy bear playing in the pool''' snake_case_ = torch.Generator(device=_UpperCAmelCase ).manual_seed(0 ) snake_case_ = pipe([prompt] , generator=_UpperCAmelCase , num_inference_steps=2 , output_type='''np''' ) snake_case_ = output.images snake_case_ = torch.Generator(device=_UpperCAmelCase ).manual_seed(0 ) snake_case_ = pipe( [prompt] , generator=_UpperCAmelCase , output_type='''np''' , return_dict=_UpperCAmelCase , num_inference_steps=2 )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) snake_case_ = np.array([0.6_551, 0.6_168, 0.5_008, 0.5_676, 0.5_659, 0.4_295, 0.6_073, 0.5_599, 0.4_992] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): snake_case_ = '''cpu''' snake_case_ = self.dummy_vqvae snake_case_ = self.dummy_text_encoder snake_case_ = self.dummy_tokenizer snake_case_ = self.dummy_transformer snake_case_ = VQDiffusionScheduler(self.num_embed ) snake_case_ = LearnedClassifierFreeSamplingEmbeddings( learnable=_UpperCAmelCase , hidden_size=self.text_embedder_hidden_size , length=tokenizer.model_max_length ) snake_case_ = VQDiffusionPipeline( vqvae=_UpperCAmelCase , text_encoder=_UpperCAmelCase , tokenizer=_UpperCAmelCase , transformer=_UpperCAmelCase , scheduler=_UpperCAmelCase , learned_classifier_free_sampling_embeddings=_UpperCAmelCase , ) snake_case_ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) snake_case_ = '''teddy bear playing in the pool''' snake_case_ = torch.Generator(device=_UpperCAmelCase ).manual_seed(0 ) snake_case_ = pipe([prompt] , generator=_UpperCAmelCase , num_inference_steps=2 , output_type='''np''' ) snake_case_ = output.images snake_case_ = torch.Generator(device=_UpperCAmelCase ).manual_seed(0 ) snake_case_ = pipe( [prompt] , generator=_UpperCAmelCase , output_type='''np''' , return_dict=_UpperCAmelCase , num_inference_steps=2 )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) snake_case_ = np.array([0.6_693, 0.6_075, 0.4_959, 0.5_701, 0.5_583, 0.4_333, 0.6_171, 0.5_684, 0.4_988] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): snake_case_ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy''' ) snake_case_ = VQDiffusionPipeline.from_pretrained('''microsoft/vq-diffusion-ithq''' ) snake_case_ = pipeline.to(_UpperCAmelCase ) pipeline.set_progress_bar_config(disable=_UpperCAmelCase ) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though snake_case_ = torch.Generator(device=_UpperCAmelCase ).manual_seed(0 ) snake_case_ = pipeline( '''teddy bear playing in the pool''' , num_images_per_prompt=1 , generator=_UpperCAmelCase , output_type='''np''' , ) snake_case_ = output.images[0] assert image.shape == (2_56, 2_56, 3) assert np.abs(expected_image - image ).max() < 2.0
267
1
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : Dict = logging.get_logger(__name__) def a__ ( A_ ): '''simple docstring''' __magic_name__ = torch.load(A_, map_location="""cpu""" ) if "model" in sd.keys(): __magic_name__ = torch.load(A_, map_location="""cpu""" )["""model"""] # pop unnecessary weights __magic_name__ = [ """decoder.version""", """decoder.output_projection.weight""", ] for key in keys_to_delete: if key in sd: sd.pop(A_ ) __magic_name__ = { """decoder.project_in_dim.weight""": """decoder.project_in.weight""", """decoder.project_out_dim.weight""": """decoder.project_out.weight""", """decoder.layer_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.layer_norm.bias""": """decoder.final_layer_norm.bias""", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __magic_name__ = sd.pop(A_ ) __magic_name__ = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __magic_name__ = sd[key] # We split QKV in separate Q,K,V __magic_name__ = key.replace(""".qkv_proj.""", """.q_proj.""" ) __magic_name__ = key.replace(""".qkv_proj.""", """.k_proj.""" ) __magic_name__ = key.replace(""".qkv_proj.""", """.v_proj.""" ) __magic_name__ = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __magic_name__ , __magic_name__ , __magic_name__ = torch.split(A_, depth // 3, dim=0 ) __magic_name__ = q __magic_name__ = k __magic_name__ = v del sd[key] return sd @torch.no_grad() def a__ ( A_, A_, A_=None ): '''simple docstring''' __magic_name__ = load_checkpoint(A_ ) if config is not None: __magic_name__ = OPTConfig.from_pretrained(A_ ) else: __magic_name__ = OPTConfig() __magic_name__ = OPTModel(A_ ).half().eval() model.load_state_dict(A_ ) # Check results Path(A_ ).mkdir(exist_ok=A_ ) model.save_pretrained(A_ ) if __name__ == "__main__": __lowerCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--fairseq_path', type=str, help=( 'path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:' ' https://huggingface.co/models?other=opt_metasq' ), ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--hf_config', default=None, type=str, help='Define HF config.') __lowerCAmelCase : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
88
from decimal import Decimal, getcontext from math import ceil, factorial def UpperCAmelCase ( lowercase ): """simple docstring""" if not isinstance(lowercase , lowercase ): raise TypeError('''Undefined for non-integers''' ) elif precision < 1: raise ValueError('''Undefined for non-natural numbers''' ) __lowercase = precision __lowercase = ceil(precision / 14 ) __lowercase = 426880 * Decimal(10005 ).sqrt() __lowercase = 1 __lowercase = 13591409 __lowercase = Decimal(lowercase ) for k in range(1 , lowercase ): __lowercase = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowercase ) ** 3) linear_term += 545140134 exponential_term *= -262537412640768000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __a : Optional[Any] = 5_0 print(F'''The first {n} digits of pi is: {pi(n)}''')
210
0
import os from collections.abc import Iterator def a ( SCREAMING_SNAKE_CASE_ : str = "." ): """simple docstring""" for dir_path, dir_names, filenames in os.walk(SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = [d for d in dir_names if d != '''scripts''' and d[0] not in '''._'''] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(SCREAMING_SNAKE_CASE_ )[1] in (".py", ".ipynb"): yield os.path.join(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ).lstrip('''./''' ) def a ( SCREAMING_SNAKE_CASE_ : Tuple ): """simple docstring""" return F"""{i * " "}*""" if i else "\n##" def a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str ): """simple docstring""" UpperCamelCase : Dict = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(SCREAMING_SNAKE_CASE_ ) or old_parts[i] != new_part) and new_part: print(F"""{md_prefix(SCREAMING_SNAKE_CASE_ )} {new_part.replace("_" , " " ).title()}""" ) return new_path def a ( SCREAMING_SNAKE_CASE_ : str = "." ): """simple docstring""" UpperCamelCase : Optional[Any] = '''''' for filepath in sorted(good_file_paths(SCREAMING_SNAKE_CASE_ ) ): UpperCamelCase , UpperCamelCase : Optional[int] = os.path.split(SCREAMING_SNAKE_CASE_ ) if filepath != old_path: UpperCamelCase : Any = print_path(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase : int = (filepath.count(os.sep ) + 1) if filepath else 0 UpperCamelCase : Union[str, Any] = F"""{filepath}/{filename}""".replace(''' ''' , '''%20''' ) UpperCamelCase : Dict = os.path.splitext(filename.replace('''_''' , ''' ''' ).title() )[0] print(F"""{md_prefix(SCREAMING_SNAKE_CASE_ )} [{filename}]({url})""" ) if __name__ == "__main__": print_directory_md(".")
315
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def a ( SCREAMING_SNAKE_CASE_ : dict ): """simple docstring""" return (data["data"], data["target"]) def a ( SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : np.ndarray ): """simple docstring""" UpperCamelCase : Optional[Any] = XGBRegressor(verbosity=0 , random_state=4_2 ) xgb.fit(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Predict target for test data UpperCamelCase : Any = xgb.predict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = predictions.reshape(len(SCREAMING_SNAKE_CASE_ ) , 1 ) return predictions def a ( ): """simple docstring""" UpperCamelCase : Tuple = fetch_california_housing() UpperCamelCase , UpperCamelCase : Tuple = data_handling(SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = train_test_split( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , test_size=0.25 , random_state=1 ) UpperCamelCase : Optional[Any] = xgboost(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Error printing print(F"""Mean Absolute Error : {mean_absolute_error(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}""" ) print(F"""Mean Square Error : {mean_squared_error(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}""" ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
315
1
"""simple docstring""" import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration SCREAMING_SNAKE_CASE_ : Dict = 5_0_0_0_0_0 SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ : Optional[Any] = os.path.split(__file__) SCREAMING_SNAKE_CASE_ : Optional[int] = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _snake_case ( UpperCAmelCase_ : datasets.Dataset , **UpperCAmelCase_ : str ): A__ = dataset.map(**UpperCAmelCase_ ) @get_duration def _snake_case ( UpperCAmelCase_ : datasets.Dataset , **UpperCAmelCase_ : Any ): A__ = dataset.filter(**UpperCAmelCase_ ) def _snake_case ( ): A__ = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: A__ = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) A__ = generate_example_dataset( os.path.join(UpperCAmelCase_ , """dataset.arrow""" ) , UpperCAmelCase_ , num_examples=UpperCAmelCase_ ) A__ = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=UpperCAmelCase_ ) def tokenize(UpperCAmelCase_ : Optional[int] ): return tokenizer(examples["""text"""] ) A__ = map(UpperCAmelCase_ ) A__ = map(UpperCAmelCase_ , batched=UpperCAmelCase_ ) A__ = map(UpperCAmelCase_ , function=lambda UpperCAmelCase_ : None , batched=UpperCAmelCase_ ) with dataset.formatted_as(type="""numpy""" ): A__ = map(UpperCAmelCase_ , function=lambda UpperCAmelCase_ : None , batched=UpperCAmelCase_ ) with dataset.formatted_as(type="""pandas""" ): A__ = map(UpperCAmelCase_ , function=lambda UpperCAmelCase_ : None , batched=UpperCAmelCase_ ) with dataset.formatted_as(type="""torch""" , columns="""numbers""" ): A__ = map(UpperCAmelCase_ , function=lambda UpperCAmelCase_ : None , batched=UpperCAmelCase_ ) with dataset.formatted_as(type="""tensorflow""" , columns="""numbers""" ): A__ = map(UpperCAmelCase_ , function=lambda UpperCAmelCase_ : None , batched=UpperCAmelCase_ ) A__ = map(UpperCAmelCase_ , function=UpperCAmelCase_ , batched=UpperCAmelCase_ ) A__ = filter(UpperCAmelCase_ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(UpperCAmelCase_ , """wb""" ) as f: f.write(json.dumps(UpperCAmelCase_ ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
335
"""simple docstring""" import math def _snake_case ( UpperCAmelCase_ : float , UpperCAmelCase_ : float ): if initial_intensity < 0: raise ValueError("""The value of intensity cannot be negative""" ) # handling of negative values of initial intensity if angle < 0 or angle > 360: raise ValueError("""In Malus Law, the angle is in the range 0-360 degrees""" ) # handling of values out of allowed range return initial_intensity * (math.cos(math.radians(UpperCAmelCase_ ) ) ** 2) if __name__ == "__main__": import doctest doctest.testmod(name='malus_law')
335
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __UpperCAmelCase : Optional[Any] = { "configuration_bloom": ["BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP", "BloomConfig", "BloomOnnxConfig"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : List[str] = ["BloomTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : Optional[int] = [ "BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST", "BloomForCausalLM", "BloomModel", "BloomPreTrainedModel", "BloomForSequenceClassification", "BloomForTokenClassification", "BloomForQuestionAnswering", ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys __UpperCAmelCase : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
293
from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = 42 class __snake_case ( __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = True @register_to_config def __init__( self : Union[str, Any] , A : int = 3 , A : int = 3 , A : Tuple[str] = ("DownEncoderBlock2D",) , A : Tuple[str] = ("UpDecoderBlock2D",) , A : Tuple[int] = (64,) , A : int = 1 , A : str = "silu" , A : int = 4 , A : int = 32 , A : int = 32 , A : float = 0.1_8215 , ): super().__init__() # pass init params to Encoder __snake_case: Any = Encoder( in_channels=A , out_channels=A , down_block_types=A , block_out_channels=A , layers_per_block=A , act_fn=A , norm_num_groups=A , double_z=A , ) # pass init params to Decoder __snake_case: int = Decoder( in_channels=A , out_channels=A , up_block_types=A , block_out_channels=A , layers_per_block=A , norm_num_groups=A , act_fn=A , ) __snake_case: Dict = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) __snake_case: int = nn.Convad(A , A , 1 ) __snake_case: List[str] = False __snake_case: Optional[int] = False # only relevant if vae tiling is enabled __snake_case: Any = self.config.sample_size __snake_case: int = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) __snake_case: Union[str, Any] = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) __snake_case: Optional[int] = 0.25 def UpperCAmelCase__ ( self : int , A : List[str] , A : Optional[Any]=False ): if isinstance(A , (Encoder, Decoder) ): __snake_case: str = value def UpperCAmelCase__ ( self : str , A : bool = True ): __snake_case: Union[str, Any] = use_tiling def UpperCAmelCase__ ( self : Optional[int] ): self.enable_tiling(A ) def UpperCAmelCase__ ( self : List[Any] ): __snake_case: List[str] = True def UpperCAmelCase__ ( self : List[str] ): __snake_case: List[str] = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def UpperCAmelCase__ ( self : Tuple ): __snake_case: Any = {} def fn_recursive_add_processors(A : str , A : torch.nn.Module , A : Dict[str, AttentionProcessor] ): if hasattr(A , """set_processor""" ): __snake_case: List[Any] = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , A , A ) return processors for name, module in self.named_children(): fn_recursive_add_processors(A , A , A ) return processors def UpperCAmelCase__ ( self : Optional[int] , A : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ): __snake_case: Any = len(self.attn_processors.keys() ) if isinstance(A , A ) and len(A ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(A )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(A : str , A : torch.nn.Module , A : Optional[Any] ): if hasattr(A , """set_processor""" ): if not isinstance(A , A ): module.set_processor(A ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , A , A ) for name, module in self.named_children(): fn_recursive_attn_processor(A , A , A ) def UpperCAmelCase__ ( self : List[str] ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def UpperCAmelCase__ ( self : Optional[Any] , A : torch.FloatTensor , A : bool = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(A , return_dict=A ) if self.use_slicing and x.shape[0] > 1: __snake_case: List[Any] = [self.encoder(A ) for x_slice in x.split(1 )] __snake_case: Optional[Any] = torch.cat(A ) else: __snake_case: str = self.encoder(A ) __snake_case: Any = self.quant_conv(A ) __snake_case: Tuple = DiagonalGaussianDistribution(A ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=A ) def UpperCAmelCase__ ( self : Tuple , A : torch.FloatTensor , A : bool = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(A , return_dict=A ) __snake_case: Optional[int] = self.post_quant_conv(A ) __snake_case: Union[str, Any] = self.decoder(A ) if not return_dict: return (dec,) return DecoderOutput(sample=A ) @apply_forward_hook def UpperCAmelCase__ ( self : Tuple , A : torch.FloatTensor , A : bool = True ): if self.use_slicing and z.shape[0] > 1: __snake_case: Union[str, Any] = [self._decode(A ).sample for z_slice in z.split(1 )] __snake_case: List[str] = torch.cat(A ) else: __snake_case: int = self._decode(A ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=A ) def UpperCAmelCase__ ( self : Any , A : Tuple , A : int , A : List[Any] ): __snake_case: int = min(a.shape[2] , b.shape[2] , A ) for y in range(A ): __snake_case: Dict = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def UpperCAmelCase__ ( self : Union[str, Any] , A : Optional[Any] , A : List[str] , A : List[str] ): __snake_case: Dict = min(a.shape[3] , b.shape[3] , A ) for x in range(A ): __snake_case: Tuple = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def UpperCAmelCase__ ( self : int , A : torch.FloatTensor , A : bool = True ): __snake_case: List[str] = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) __snake_case: Dict = int(self.tile_latent_min_size * self.tile_overlap_factor ) __snake_case: Dict = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. __snake_case: Optional[int] = [] for i in range(0 , x.shape[2] , A ): __snake_case: Optional[int] = [] for j in range(0 , x.shape[3] , A ): __snake_case: int = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] __snake_case: Tuple = self.encoder(A ) __snake_case: Dict = self.quant_conv(A ) row.append(A ) rows.append(A ) __snake_case: Tuple = [] for i, row in enumerate(A ): __snake_case: str = [] for j, tile in enumerate(A ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __snake_case: Optional[Any] = self.blend_v(rows[i - 1][j] , A , A ) if j > 0: __snake_case: Tuple = self.blend_h(row[j - 1] , A , A ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(A , dim=3 ) ) __snake_case: Tuple = torch.cat(A , dim=2 ) __snake_case: Optional[int] = DiagonalGaussianDistribution(A ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=A ) def UpperCAmelCase__ ( self : Union[str, Any] , A : torch.FloatTensor , A : bool = True ): __snake_case: Optional[Any] = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) __snake_case: str = int(self.tile_sample_min_size * self.tile_overlap_factor ) __snake_case: int = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. __snake_case: List[Any] = [] for i in range(0 , z.shape[2] , A ): __snake_case: Optional[Any] = [] for j in range(0 , z.shape[3] , A ): __snake_case: Dict = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] __snake_case: Any = self.post_quant_conv(A ) __snake_case: Optional[Any] = self.decoder(A ) row.append(A ) rows.append(A ) __snake_case: Optional[Any] = [] for i, row in enumerate(A ): __snake_case: Optional[Any] = [] for j, tile in enumerate(A ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __snake_case: Tuple = self.blend_v(rows[i - 1][j] , A , A ) if j > 0: __snake_case: List[str] = self.blend_h(row[j - 1] , A , A ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(A , dim=3 ) ) __snake_case: Dict = torch.cat(A , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=A ) def UpperCAmelCase__ ( self : List[Any] , A : torch.FloatTensor , A : bool = False , A : bool = True , A : Optional[torch.Generator] = None , ): __snake_case: Optional[Any] = sample __snake_case: Union[str, Any] = self.encode(A ).latent_dist if sample_posterior: __snake_case: Optional[Any] = posterior.sample(generator=A ) else: __snake_case: Dict = posterior.mode() __snake_case: Any = self.decode(A ).sample if not return_dict: return (dec,) return DecoderOutput(sample=A )
293
1
def _a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : list[str] ): __lowerCAmelCase = "" for word_or_phrase in separated: if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise Exception("join() accepts only strings to be joined" ) joined += word_or_phrase + separator return joined.strip(lowerCamelCase__ ) if __name__ == "__main__": from doctest import testmod testmod()
92
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowercase ( a_ ): """simple docstring""" UpperCamelCase : Any = ["image_processor", "tokenizer"] UpperCamelCase : Dict = "BridgeTowerImageProcessor" UpperCamelCase : List[Any] = ("RobertaTokenizer", "RobertaTokenizerFast") def __init__( self , A , A ) -> Optional[int]: '''simple docstring''' super().__init__(A , A ) def __call__( self , A , A = None , A = True , A = False , A = None , A = None , A = 0 , A = None , A = None , A = None , A = False , A = False , A = False , A = False , A = True , A = None , **A , ) -> BatchEncoding: '''simple docstring''' lowerCamelCase = self.tokenizer( text=A , add_special_tokens=A , padding=A , truncation=A , max_length=A , stride=A , pad_to_multiple_of=A , return_token_type_ids=A , return_attention_mask=A , return_overflowing_tokens=A , return_special_tokens_mask=A , return_offsets_mapping=A , return_length=A , verbose=A , return_tensors=A , **A , ) # add pixel_values + pixel_mask lowerCamelCase = self.image_processor( A , return_tensors=A , do_normalize=A , do_center_crop=A , **A ) encoding.update(A ) return encoding def __A ( self , *A , **A ) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*A , **A ) def __A ( self , *A , **A ) -> Optional[int]: '''simple docstring''' return self.tokenizer.decode(*A , **A ) @property def __A ( self ) -> Dict: '''simple docstring''' lowerCamelCase = self.tokenizer.model_input_names lowerCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
252
0
import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , UpperCamelCase_ , UpperCamelCase_=7 , UpperCamelCase_=3 , UpperCamelCase_=18 , UpperCamelCase_=30 , UpperCamelCase_=400 , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=True , UpperCamelCase_=[0.5, 0.5, 0.5] , UpperCamelCase_=[0.5, 0.5, 0.5] , UpperCamelCase_=False , ): lowercase_ :Optional[Any] = size if size is not None else {'''height''': 20, '''width''': 20} lowercase_ :int = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} lowercase_ :List[str] = parent lowercase_ :List[Any] = batch_size lowercase_ :Tuple = num_channels lowercase_ :str = image_size lowercase_ :Optional[Any] = min_resolution lowercase_ :int = max_resolution lowercase_ :str = do_resize lowercase_ :List[str] = size lowercase_ :List[str] = do_center_crop lowercase_ :Optional[int] = crop_size lowercase_ :Optional[Any] = do_normalize lowercase_ :int = image_mean lowercase_ :Optional[int] = image_std lowercase_ :Dict = do_reduce_labels def UpperCamelCase ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCamelCase ( ) -> Optional[int]: '''simple docstring''' lowercase_ :int = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) lowercase_ :List[str] = Image.open(dataset[0]['''file'''] ) lowercase_ :List[str] = Image.open(dataset[1]['''file'''] ) return image, map def UpperCamelCase ( ) -> Tuple: '''simple docstring''' lowercase_ :Dict = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) lowercase_ :Optional[Any] = Image.open(ds[0]['''file'''] ) lowercase_ :List[str] = Image.open(ds[1]['''file'''] ) lowercase_ :List[Any] = Image.open(ds[2]['''file'''] ) lowercase_ :Any = Image.open(ds[3]['''file'''] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class UpperCamelCase ( lowercase__ , unittest.TestCase ): '''simple docstring''' lowercase : Optional[Any] =BeitImageProcessor if is_vision_available() else None def UpperCamelCase ( self ): lowercase_ :Dict = BeitImageProcessingTester(self ) @property def UpperCamelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase ( self ): lowercase_ :Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''size''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''image_std''' ) ) def UpperCamelCase ( self ): lowercase_ :List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 20, '''width''': 20} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) self.assertEqual(image_processor.do_reduce_labels , UpperCamelCase_ ) lowercase_ :Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=UpperCamelCase_ ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) self.assertEqual(image_processor.do_reduce_labels , UpperCamelCase_ ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # Initialize image_processing lowercase_ :Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ :List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image ) # Test not batched input lowercase_ :Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase_ :List[str] = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase ( self ): # Initialize image_processing lowercase_ :Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ :Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray ) # Test not batched input lowercase_ :Union[str, Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase_ :List[str] = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase ( self ): # Initialize image_processing lowercase_ :Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ :List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor ) # Test not batched input lowercase_ :List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase_ :str = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase ( self ): # Initialize image_processing lowercase_ :Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ :Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_ ) lowercase_ :Tuple = [] for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input lowercase_ :Tuple = image_processing(image_inputs[0] , maps[0] , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 1, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 255 ) # Test batched lowercase_ :Optional[int] = image_processing(UpperCamelCase_ , UpperCamelCase_ , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 255 ) # Test not batched input (PIL images) lowercase_ :List[Any] = prepare_semantic_single_inputs() lowercase_ :Union[str, Any] = image_processing(UpperCamelCase_ , UpperCamelCase_ , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 1, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 255 ) # Test batched input (PIL images) lowercase_ :Tuple = prepare_semantic_batch_inputs() lowercase_ :str = image_processing(UpperCamelCase_ , UpperCamelCase_ , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 2, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 255 ) def UpperCamelCase ( self ): # Initialize image_processing lowercase_ :Tuple = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 lowercase_ :Tuple = prepare_semantic_single_inputs() lowercase_ :Dict = image_processing(UpperCamelCase_ , UpperCamelCase_ , return_tensors='''pt''' ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 150 ) lowercase_ :Dict = True lowercase_ :Tuple = image_processing(UpperCamelCase_ , UpperCamelCase_ , return_tensors='''pt''' ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 255 )
352
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Dict = { "bigcode/gpt_bigcode-santacoder": "https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json", } class UpperCamelCase ( lowercase__ ): '''simple docstring''' lowercase : List[Any] ="""gpt_bigcode""" lowercase : Dict =["""past_key_values"""] lowercase : List[Any] ={ """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , UpperCamelCase_=5_0257 , UpperCamelCase_=1024 , UpperCamelCase_=768 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=None , UpperCamelCase_="gelu_pytorch_tanh" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=1E-5 , UpperCamelCase_=0.02 , UpperCamelCase_=True , UpperCamelCase_=True , UpperCamelCase_=5_0256 , UpperCamelCase_=5_0256 , UpperCamelCase_=True , UpperCamelCase_=True , UpperCamelCase_=True , **UpperCamelCase_ , ): lowercase_ :Any = vocab_size lowercase_ :List[str] = n_positions lowercase_ :Union[str, Any] = n_embd lowercase_ :Dict = n_layer lowercase_ :Optional[int] = n_head lowercase_ :List[str] = n_inner lowercase_ :List[str] = activation_function lowercase_ :Optional[int] = resid_pdrop lowercase_ :Union[str, Any] = embd_pdrop lowercase_ :Any = attn_pdrop lowercase_ :Optional[Any] = layer_norm_epsilon lowercase_ :str = initializer_range lowercase_ :Optional[Any] = scale_attn_weights lowercase_ :Any = use_cache lowercase_ :Union[str, Any] = attention_softmax_in_fpaa lowercase_ :int = scale_attention_softmax_in_fpaa lowercase_ :Union[str, Any] = multi_query lowercase_ :List[str] = bos_token_id lowercase_ :Optional[int] = eos_token_id super().__init__(bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ )
252
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule UpperCamelCase_ = {'tokenization_byt5': ['ByT5Tokenizer']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
251
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
0
def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase )-> int: '''simple docstring''' UpperCAmelCase : List[Any] =1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): UpperCAmelCase : Tuple =n - k # Calculate C(n,k) for i in range(__lowerCAmelCase ): result *= n - i result //= i + 1 return result def lowerCAmelCase_ ( __lowerCAmelCase )-> int: '''simple docstring''' return binomial_coefficient(2 * node_count , __lowerCAmelCase ) // (node_count + 1) def lowerCAmelCase_ ( __lowerCAmelCase )-> int: '''simple docstring''' if n < 0: raise ValueError('''factorial() not defined for negative values''' ) UpperCAmelCase : Any =1 for i in range(1 , n + 1 ): result *= i return result def lowerCAmelCase_ ( __lowerCAmelCase )-> int: '''simple docstring''' return catalan_number(__lowerCAmelCase ) * factorial(__lowerCAmelCase ) if __name__ == "__main__": __snake_case = int(input('''Enter the number of nodes: ''').strip() or 0) if node_count <= 0: raise ValueError('''We need some nodes to work with.''') print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
357
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge __snake_case = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] __snake_case = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def lowerCAmelCase_ ( )-> Optional[Any]: '''simple docstring''' UpperCAmelCase : Optional[int] =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , bootstrap_aggregation=__lowerCAmelCase , rouge_keys=['''rouge2''', '''rougeL'''] ) assert isinstance(__lowerCAmelCase , __lowerCAmelCase ) UpperCAmelCase : List[Any] =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , bootstrap_aggregation=__lowerCAmelCase , rouge_keys=['''rouge2'''] ) assert ( pd.DataFrame(no_aggregation['''rouge2'''] ).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra['''rouge2'''] ).fmeasure.mean() ) def lowerCAmelCase_ ( )-> Dict: '''simple docstring''' UpperCAmelCase : Any ='''rougeLsum''' UpperCAmelCase : Optional[Any] =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase , rouge_keys=[k] )[k] UpperCAmelCase : List[Any] =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase , rouge_keys=[k] )[k] assert score > score_no_sep def lowerCAmelCase_ ( )-> Any: '''simple docstring''' UpperCAmelCase : str =['''rouge1''', '''rouge2''', '''rougeL'''] UpperCAmelCase : int =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase , rouge_keys=__lowerCAmelCase ) UpperCAmelCase : Tuple =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase , rouge_keys=__lowerCAmelCase ) assert score_sep == score_no_sep def lowerCAmelCase_ ( )-> Dict: '''simple docstring''' UpperCAmelCase : int =[ '''Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.''', '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''', ] UpperCAmelCase : Any =[ '''Margot Frank, died in 1945, a month earlier than previously thought.''', '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of''' ''' the final seconds on board Flight 9525.''', ] assert calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase ) == calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , newline_sep=__lowerCAmelCase ) def lowerCAmelCase_ ( )-> List[str]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =[ '''" "a person who has such a video needs to immediately give it to the investigators," prosecutor says .<n> "it is a very disturbing scene," editor-in-chief of bild online tells "erin burnett: outfront" ''' ] UpperCAmelCase : Optional[Any] =[ ''' Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports . Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says .''' ] UpperCAmelCase : Optional[int] =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , rouge_keys=['''rougeLsum'''] , newline_sep=__lowerCAmelCase )['''rougeLsum'''] UpperCAmelCase : int =calculate_rouge(__lowerCAmelCase , __lowerCAmelCase , rouge_keys=['''rougeLsum'''] )['''rougeLsum'''] assert new_score > prev_score def lowerCAmelCase_ ( )-> Optional[int]: '''simple docstring''' UpperCAmelCase : List[Any] =Path('''examples/seq2seq/test_data/wmt_en_ro''' ) UpperCAmelCase : Tuple =calculate_rouge_path(data_dir.joinpath('''test.source''' ) , data_dir.joinpath('''test.target''' ) ) assert isinstance(__lowerCAmelCase , __lowerCAmelCase ) UpperCAmelCase : Dict =calculate_rouge_path( data_dir.joinpath('''test.source''' ) , data_dir.joinpath('''test.target''' ) , bootstrap_aggregation=__lowerCAmelCase ) assert isinstance(__lowerCAmelCase , __lowerCAmelCase )
78
0
"""simple docstring""" # flake8: noqa # Lint as: python3 lowerCamelCase__ = [ """VerificationMode""", """Version""", """disable_progress_bar""", """enable_progress_bar""", """is_progress_bar_enabled""", """experimental""", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
86
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class A__ ( _lowerCamelCase , unittest.TestCase): A_ : Union[str, Any] = BarthezTokenizer A_ : Tuple = BarthezTokenizerFast A_ : Dict = True A_ : List[str] = True def __lowerCamelCase ( self ): super().setUp() __lowerCAmelCase : str = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Any = tokenizer def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = '<pad>' __lowerCAmelCase : Union[str, Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): __lowerCAmelCase : List[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , 10_11_22 ) def __lowerCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def __lowerCamelCase ( self ): __lowerCAmelCase : List[str] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] __lowerCAmelCase : Optional[Any] = [0, 57, 30_18, 7_03_07, 91, 2] __lowerCAmelCase : Optional[int] = self.tokenizer( _SCREAMING_SNAKE_CASE , max_length=len(_SCREAMING_SNAKE_CASE ) , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __lowerCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): if not self.test_rust_tokenizer: return __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : Optional[int] = self.get_rust_tokenizer() __lowerCAmelCase : List[str] = 'I was born in 92000, and this is falsé.' __lowerCAmelCase : Optional[int] = tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = rust_tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = self.get_rust_tokenizer() __lowerCAmelCase : List[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Dict = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): # fmt: off __lowerCAmelCase : str = {'input_ids': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __lowerCAmelCase : Union[str, Any] = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_SCREAMING_SNAKE_CASE , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=_SCREAMING_SNAKE_CASE , )
86
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = 1 __snake_case : Optional[int] = 3 __snake_case : Dict = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(a_ ) return image @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' torch.manual_seed(0 ) __snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' torch.manual_seed(0 ) __snake_case : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' torch.manual_seed(0 ) __snake_case : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=50_06 , ) return RobertaSeriesModelWithTransformation(a_ ) @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' def extract(*a_ , **a_ ): class _UpperCAmelCase : '''simple docstring''' def __init__(self ): '''simple docstring''' __snake_case : Optional[Any] = torch.ones([0] ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' self.pixel_values.to(a_ ) return self return Out() return extract def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __snake_case : Dict = self.dummy_cond_unet __snake_case : int = PNDMScheduler(skip_prk_steps=a_ ) __snake_case : Tuple = self.dummy_vae __snake_case : Optional[Any] = self.dummy_text_encoder __snake_case : Dict = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) __snake_case : List[str] = 77 __snake_case : List[str] = self.dummy_image.to(a_ ) __snake_case : Dict = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk __snake_case : Union[str, Any] = AltDiffusionImgaImgPipeline( unet=a_ , scheduler=a_ , vae=a_ , text_encoder=a_ , tokenizer=a_ , safety_checker=a_ , feature_extractor=self.dummy_extractor , ) __snake_case : str = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=a_ ) __snake_case : Dict = alt_pipe.to(a_ ) alt_pipe.set_progress_bar_config(disable=a_ ) __snake_case : int = '''A painting of a squirrel eating a burger''' __snake_case : Tuple = torch.Generator(device=a_ ).manual_seed(0 ) __snake_case : Dict = alt_pipe( [prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=a_ , ) __snake_case : List[Any] = output.images __snake_case : List[Any] = torch.Generator(device=a_ ).manual_seed(0 ) __snake_case : Tuple = alt_pipe( [prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=a_ , return_dict=a_ , )[0] __snake_case : Union[str, Any] = image[0, -3:, -3:, -1] __snake_case : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __snake_case : Dict = np.array([0.4427, 0.3731, 0.4249, 0.4941, 0.4546, 0.4148, 0.4193, 0.4666, 0.4499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[int] = self.dummy_cond_unet __snake_case : Optional[Any] = PNDMScheduler(skip_prk_steps=a_ ) __snake_case : Any = self.dummy_vae __snake_case : List[Any] = self.dummy_text_encoder __snake_case : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) __snake_case : Tuple = 77 __snake_case : Optional[Any] = self.dummy_image.to(a_ ) # put models in fp16 __snake_case : List[Any] = unet.half() __snake_case : int = vae.half() __snake_case : Tuple = bert.half() # make sure here that pndm scheduler skips prk __snake_case : Optional[Any] = AltDiffusionImgaImgPipeline( unet=a_ , scheduler=a_ , vae=a_ , text_encoder=a_ , tokenizer=a_ , safety_checker=a_ , feature_extractor=self.dummy_extractor , ) __snake_case : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=a_ ) __snake_case : int = alt_pipe.to(a_ ) alt_pipe.set_progress_bar_config(disable=a_ ) __snake_case : Optional[Any] = '''A painting of a squirrel eating a burger''' __snake_case : Dict = torch.manual_seed(0 ) __snake_case : Any = alt_pipe( [prompt] , generator=a_ , num_inference_steps=2 , output_type='''np''' , image=a_ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) # resize to resolution that is divisible by 8 but not 16 or 32 __snake_case : Optional[Any] = init_image.resize((7_60, 5_04) ) __snake_case : Optional[Any] = '''BAAI/AltDiffusion''' __snake_case : Tuple = AltDiffusionImgaImgPipeline.from_pretrained( a_ , safety_checker=a_ , ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) pipe.enable_attention_slicing() __snake_case : List[Any] = '''A fantasy landscape, trending on artstation''' __snake_case : Dict = torch.manual_seed(0 ) __snake_case : Optional[Any] = pipe( prompt=a_ , image=a_ , strength=0.75 , guidance_scale=7.5 , generator=a_ , output_type='''np''' , ) __snake_case : Union[str, Any] = output.images[0] __snake_case : Any = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 7_60, 3) __snake_case : Dict = np.array([0.9358, 0.9397, 0.9599, 0.9901, 1.0000, 1.0000, 0.9882, 1.0000, 1.0000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __snake_case : Dict = init_image.resize((7_68, 5_12) ) __snake_case : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''' ) __snake_case : Tuple = '''BAAI/AltDiffusion''' __snake_case : Optional[int] = AltDiffusionImgaImgPipeline.from_pretrained( a_ , safety_checker=a_ , ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) pipe.enable_attention_slicing() __snake_case : Optional[Any] = '''A fantasy landscape, trending on artstation''' __snake_case : Optional[Any] = torch.manual_seed(0 ) __snake_case : str = pipe( prompt=a_ , image=a_ , strength=0.75 , guidance_scale=7.5 , generator=a_ , output_type='''np''' , ) __snake_case : List[str] = output.images[0] assert image.shape == (5_12, 7_68, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
357
"""simple docstring""" import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def lowercase ( _snake_case : str , _snake_case : str , _snake_case : str ) ->List[Any]: """simple docstring""" def get_masked_lm_array(_snake_case : str ): __snake_case : int = f"""masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE""" __snake_case : str = tf.train.load_variable(_snake_case , _snake_case ) if "kernel" in name: __snake_case : Any = array.transpose() return torch.from_numpy(_snake_case ) def get_encoder_array(_snake_case : str ): __snake_case : List[str] = f"""encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE""" __snake_case : Union[str, Any] = tf.train.load_variable(_snake_case , _snake_case ) if "kernel" in name: __snake_case : Optional[int] = array.transpose() return torch.from_numpy(_snake_case ) def get_encoder_layer_array(_snake_case : int , _snake_case : str ): __snake_case : str = f"""encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE""" __snake_case : Optional[int] = tf.train.load_variable(_snake_case , _snake_case ) if "kernel" in name: __snake_case : Optional[Any] = array.transpose() return torch.from_numpy(_snake_case ) def get_encoder_attention_layer_array(_snake_case : int , _snake_case : str , _snake_case : str ): __snake_case : Any = f"""encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE""" __snake_case : Dict = tf.train.load_variable(_snake_case , _snake_case ) __snake_case : int = array.reshape(_snake_case ) if "kernel" in name: __snake_case : Optional[int] = array.transpose() return torch.from_numpy(_snake_case ) print(f"""Loading model based on config from {config_path}...""" ) __snake_case : Optional[Any] = BertConfig.from_json_file(_snake_case ) __snake_case : Dict = BertForMaskedLM(_snake_case ) # Layers for layer_index in range(0 , config.num_hidden_layers ): __snake_case : BertLayer = model.bert.encoder.layer[layer_index] # Self-attention __snake_case : BertSelfAttention = layer.attention.self __snake_case : int = get_encoder_attention_layer_array( _snake_case , '''_query_dense/kernel''' , self_attn.query.weight.data.shape ) __snake_case : str = get_encoder_attention_layer_array( _snake_case , '''_query_dense/bias''' , self_attn.query.bias.data.shape ) __snake_case : str = get_encoder_attention_layer_array( _snake_case , '''_key_dense/kernel''' , self_attn.key.weight.data.shape ) __snake_case : List[Any] = get_encoder_attention_layer_array( _snake_case , '''_key_dense/bias''' , self_attn.key.bias.data.shape ) __snake_case : Tuple = get_encoder_attention_layer_array( _snake_case , '''_value_dense/kernel''' , self_attn.value.weight.data.shape ) __snake_case : Union[str, Any] = get_encoder_attention_layer_array( _snake_case , '''_value_dense/bias''' , self_attn.value.bias.data.shape ) # Self-attention Output __snake_case : BertSelfOutput = layer.attention.output __snake_case : Dict = get_encoder_attention_layer_array( _snake_case , '''_output_dense/kernel''' , self_output.dense.weight.data.shape ) __snake_case : Tuple = get_encoder_attention_layer_array( _snake_case , '''_output_dense/bias''' , self_output.dense.bias.data.shape ) __snake_case : str = get_encoder_layer_array(_snake_case , '''_attention_layer_norm/gamma''' ) __snake_case : Any = get_encoder_layer_array(_snake_case , '''_attention_layer_norm/beta''' ) # Intermediate __snake_case : BertIntermediate = layer.intermediate __snake_case : int = get_encoder_layer_array(_snake_case , '''_intermediate_dense/kernel''' ) __snake_case : int = get_encoder_layer_array(_snake_case , '''_intermediate_dense/bias''' ) # Output __snake_case : BertOutput = layer.output __snake_case : List[str] = get_encoder_layer_array(_snake_case , '''_output_dense/kernel''' ) __snake_case : Dict = get_encoder_layer_array(_snake_case , '''_output_dense/bias''' ) __snake_case : List[str] = get_encoder_layer_array(_snake_case , '''_output_layer_norm/gamma''' ) __snake_case : Union[str, Any] = get_encoder_layer_array(_snake_case , '''_output_layer_norm/beta''' ) # Embeddings __snake_case : Optional[int] = get_encoder_array('''_position_embedding_layer/embeddings''' ) __snake_case : str = get_encoder_array('''_type_embedding_layer/embeddings''' ) __snake_case : int = get_encoder_array('''_embedding_norm_layer/gamma''' ) __snake_case : Tuple = get_encoder_array('''_embedding_norm_layer/beta''' ) # LM Head __snake_case : Optional[Any] = model.cls.predictions.transform __snake_case : Dict = get_masked_lm_array('''dense/kernel''' ) __snake_case : Union[str, Any] = get_masked_lm_array('''dense/bias''' ) __snake_case : str = get_masked_lm_array('''layer_norm/gamma''' ) __snake_case : Tuple = get_masked_lm_array('''layer_norm/beta''' ) __snake_case : Tuple = get_masked_lm_array('''embedding_table''' ) # Pooling __snake_case : Optional[Any] = BertPooler(config=_snake_case ) __snake_case : BertPooler = get_encoder_array('''_pooler_layer/kernel''' ) __snake_case : BertPooler = get_encoder_array('''_pooler_layer/bias''' ) # Export final model model.save_pretrained(_snake_case ) # Integration test - should load without any errors ;) __snake_case : Dict = BertForMaskedLM.from_pretrained(_snake_case ) print(new_model.eval() ) print('''Model conversion was done sucessfully!''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() parser.add_argument( """--tf_checkpoint_path""", type=str, required=True, help="""Path to the TensorFlow Token Dropping checkpoint path.""" ) parser.add_argument( """--bert_config_file""", type=str, required=True, help="""The config json file corresponding to the BERT model. This specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", type=str, required=True, help="""Path to the output PyTorch model.""", ) SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
24
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase__ = { 'vocab_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/vocab.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/vocab.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/vocab.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/vocab.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/vocab.json', }, 'merges_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/merges.txt', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/merges.txt', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/merges.txt', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/merges.txt', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/merges.txt', }, 'tokenizer_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/tokenizer.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/tokenizer.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/tokenizer.json', }, } lowerCAmelCase__ = { 'gpt2': 10_24, 'gpt2-medium': 10_24, 'gpt2-large': 10_24, 'gpt2-xl': 10_24, 'distilgpt2': 10_24, } class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] __SCREAMING_SNAKE_CASE = GPTaTokenizer def __init__( self , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase="<|endoftext|>" , __lowerCamelCase="<|endoftext|>" , __lowerCamelCase="<|endoftext|>" , __lowerCamelCase=False , **__lowerCamelCase , ) -> List[str]: super().__init__( __lowerCamelCase , __lowerCamelCase , tokenizer_file=__lowerCamelCase , unk_token=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , add_prefix_space=__lowerCamelCase , **__lowerCamelCase , ) _A : Union[str, Any] = kwargs.pop("add_bos_token" , __lowerCamelCase) _A : Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , __lowerCamelCase) != add_prefix_space: _A : int = getattr(__lowerCamelCase , pre_tok_state.pop("type")) _A : Union[str, Any] = add_prefix_space _A : Union[str, Any] = pre_tok_class(**__lowerCamelCase) _A : int = add_prefix_space def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> BatchEncoding: _A : Any = kwargs.get("is_split_into_words" , __lowerCamelCase) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> BatchEncoding: _A : Dict = kwargs.get("is_split_into_words" , __lowerCamelCase) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> Tuple[str]: _A : Optional[Any] = self._tokenizer.model.save(__lowerCamelCase , name=__lowerCamelCase) return tuple(__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> List[int]: _A : List[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase) + [self.eos_token_id]) if len(__lowerCamelCase) > self.model_max_length: _A : Any = input_ids[-self.model_max_length :] return input_ids
11
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = ["image_processor", "tokenizer"] __SCREAMING_SNAKE_CASE = "OwlViTImageProcessor" __SCREAMING_SNAKE_CASE = ("CLIPTokenizer", "CLIPTokenizerFast") def __init__( self , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase) -> Union[str, Any]: _A : int = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __lowerCamelCase , ) _A : List[Any] = kwargs.pop("feature_extractor") _A : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`.") if tokenizer is None: raise ValueError("You need to specify a `tokenizer`.") super().__init__(__lowerCamelCase , __lowerCamelCase) def __call__( self , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase="max_length" , __lowerCamelCase="np" , **__lowerCamelCase) -> Any: if text is None and query_images is None and images is None: raise ValueError( "You have to specify at least one text or query image or image. All three cannot be none.") if text is not None: if isinstance(__lowerCamelCase , __lowerCamelCase) or (isinstance(__lowerCamelCase , __lowerCamelCase) and not isinstance(text[0] , __lowerCamelCase)): _A : Union[str, Any] = [self.tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase)] elif isinstance(__lowerCamelCase , __lowerCamelCase) and isinstance(text[0] , __lowerCamelCase): _A : Optional[Any] = [] # Maximum number of queries across batch _A : str = max([len(__lowerCamelCase) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__lowerCamelCase) != max_num_queries: _A : Optional[int] = t + [" "] * (max_num_queries - len(__lowerCamelCase)) _A : List[Any] = self.tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase) encodings.append(__lowerCamelCase) else: raise TypeError("Input text should be a string, a list of strings or a nested list of strings") if return_tensors == "np": _A : Tuple = np.concatenate([encoding["input_ids"] for encoding in encodings] , axis=0) _A : Optional[Any] = np.concatenate([encoding["attention_mask"] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _A : Optional[int] = jnp.concatenate([encoding["input_ids"] for encoding in encodings] , axis=0) _A : Optional[int] = jnp.concatenate([encoding["attention_mask"] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _A : Optional[Any] = torch.cat([encoding["input_ids"] for encoding in encodings] , dim=0) _A : Union[str, Any] = torch.cat([encoding["attention_mask"] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _A : Any = tf.stack([encoding["input_ids"] for encoding in encodings] , axis=0) _A : Tuple = tf.stack([encoding["attention_mask"] for encoding in encodings] , axis=0) else: raise ValueError("Target return tensor type could not be returned") _A : Optional[Any] = BatchEncoding() _A : Tuple = input_ids _A : Dict = attention_mask if query_images is not None: _A : Optional[Any] = BatchEncoding() _A : List[str] = self.image_processor( __lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase).pixel_values _A : Union[str, Any] = query_pixel_values if images is not None: _A : int = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase) if text is not None and images is not None: _A : Tuple = image_features.pixel_values return encoding elif query_images is not None and images is not None: _A : int = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase) , tensor_type=__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> str: return self.image_processor.post_process(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> List[str]: return self.image_processor.post_process_object_detection(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> Optional[int]: return self.image_processor.post_process_image_guided_detection(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> int: return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> Optional[int]: return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase) @property def _lowerCamelCase ( self) -> int: warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __lowerCamelCase , ) return self.image_processor_class @property def _lowerCamelCase ( self) -> List[str]: warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __lowerCamelCase , ) return self.image_processor
11
1
"""simple docstring""" from math import isqrt def UpperCamelCase ( UpperCAmelCase ) ->list[int]: """simple docstring""" a_ = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , UpperCAmelCase , UpperCAmelCase ): a_ = False return [i for i in range(2 , UpperCAmelCase ) if is_prime[i]] def UpperCamelCase ( UpperCAmelCase = 10**8 ) ->int: """simple docstring""" a_ = calculate_prime_numbers(max_number // 2 ) a_ = 0 a_ = 0 a_ = len(UpperCAmelCase ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(F"""{solution() = }""")
303
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) ->tuple[float | int, list[tuple[int, int]]]: """simple docstring""" a_ , a_ = grid.shape a_ = [-1, 1, 0, 0] a_ = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] a_ , a_ = [(0, source)], set() a_ = np.full((rows, cols) , np.inf ) a_ = 0 a_ = np.empty((rows, cols) , dtype=UpperCAmelCase ) a_ = None while queue: ((a_) , (a_)) = heappop(UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: a_ = [] while (x, y) != source: path.append((x, y) ) a_ , a_ = predecessors[x, y] path.append(UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCAmelCase ) ): a_ , a_ = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: a_ = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCAmelCase , (dist + 1, (nx, ny)) ) a_ = dist + 1 a_ = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
303
1
'''simple docstring''' import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler a_ : Optional[Any] = 1_6 a_ : List[str] = 3_2 def _A (lowerCAmelCase__ :Optional[Any] ) -> str: '''simple docstring''' return int(x / 2**20 ) class a : def __enter__( self ) -> Optional[Any]: gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero _a = torch.cuda.memory_allocated() return self def __exit__( self , *__magic_name__ ) -> Any: gc.collect() torch.cuda.empty_cache() _a = torch.cuda.memory_allocated() _a = torch.cuda.max_memory_allocated() _a = bamb(self.end - self.begin ) _a = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def _A (lowerCAmelCase__ :Accelerator , lowerCAmelCase__ :int = 16 , lowerCAmelCase__ :str = "bert-base-cased" , lowerCAmelCase__ :int = 3_20 , lowerCAmelCase__ :int = 1_60 , ) -> Optional[int]: '''simple docstring''' _a = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) _a = load_dataset( 'glue' , 'mrpc' , split={'train': f'train[:{n_train}]', 'validation': f'validation[:{n_val}]'} ) def tokenize_function(lowerCAmelCase__ :int ): # max_length=None => use the model max length (it's actually the default) _a = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _a = datasets.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowerCAmelCase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowerCAmelCase__ :Optional[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase__ , padding='max_length' , max_length=1_28 , return_tensors='pt' ) return tokenizer.pad(lowerCAmelCase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _a = DataLoader( tokenized_datasets['train'] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) _a = DataLoader( tokenized_datasets['validation'] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) return train_dataloader, eval_dataloader def _A (lowerCAmelCase__ :str , lowerCAmelCase__ :List[Any] ) -> List[str]: '''simple docstring''' _a = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a = config['lr'] _a = int(config['num_epochs'] ) _a = int(config['seed'] ) _a = int(config['batch_size'] ) _a = args.model_name_or_path set_seed(lowerCAmelCase__ ) _a , _a = get_dataloaders(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , args.n_train , args.n_val ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a = AutoModelForSequenceClassification.from_pretrained(lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) # Instantiate optimizer _a = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _a = optimizer_cls(params=model.parameters() , lr=lowerCAmelCase__ ) if accelerator.state.deepspeed_plugin is not None: _a = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _a = 1 _a = (len(lowerCAmelCase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _a = get_linear_schedule_with_warmup( optimizer=lowerCAmelCase__ , num_warmup_steps=0 , num_training_steps=lowerCAmelCase__ , ) else: _a = DummyScheduler(lowerCAmelCase__ , total_num_steps=lowerCAmelCase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a , _a , _a , _a , _a = accelerator.prepare( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # We need to keep track of how many total steps we have iterated over _a = 0 # We also need to keep track of the stating epoch so files are named properly _a = 0 # Now we train the model _a = {} for epoch in range(lowerCAmelCase__ , lowerCAmelCase__ ): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(lowerCAmelCase__ ): _a = model(**lowerCAmelCase__ ) _a = outputs.loss _a = loss / gradient_accumulation_steps accelerator.backward(lowerCAmelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print('Memory before entering the train : {}'.format(bamb(tracemalloc.begin ) ) ) accelerator.print('Memory consumed at the end of the train (end-begin): {}'.format(tracemalloc.used ) ) accelerator.print('Peak Memory consumed during the train (max-begin): {}'.format(tracemalloc.peaked ) ) accelerator.print( 'Total Peak Memory consumed during the train (max): {}'.format( tracemalloc.peaked + bamb(tracemalloc.begin ) ) ) _a = tracemalloc.peaked + bamb(tracemalloc.begin ) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[f'epoch-{epoch}'] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'peak_memory_utilization.json' ) , 'w' ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) def _A () -> List[str]: '''simple docstring''' _a = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowerCAmelCase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowerCAmelCase__ , ) parser.add_argument( '--output_dir' , type=lowerCAmelCase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--peak_memory_upper_bound' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.' , ) parser.add_argument( '--n_train' , type=lowerCAmelCase__ , default=3_20 , help='Number of training examples to use.' , ) parser.add_argument( '--n_val' , type=lowerCAmelCase__ , default=1_60 , help='Number of validation examples to use.' , ) parser.add_argument( '--num_epochs' , type=lowerCAmelCase__ , default=1 , help='Number of train epochs.' , ) _a = parser.parse_args() _a = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": main()
168
'''simple docstring''' from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
168
1
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __a = { "configuration_gpt_neox_japanese": ["GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoXJapaneseConfig"], "tokenization_gpt_neox_japanese": ["GPTNeoXJapaneseTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoXJapaneseForCausalLM", "GPTNeoXJapaneseLayer", "GPTNeoXJapaneseModel", "GPTNeoXJapanesePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
43
'''simple docstring''' import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __a = logging.getLogger(__name__) class UpperCAmelCase_ ( _a ): """simple docstring""" def __init__( self : Union[str, Any] , snake_case_ : Any , snake_case_ : int , snake_case_ : str , snake_case_ : Tuple=None ): super().__init__( snake_case_ , question_encoder_tokenizer=snake_case_ , generator_tokenizer=snake_case_ , index=snake_case_ , init_retrieval=snake_case_ , ) snake_case__ : int = None def lowerCamelCase ( self : int , snake_case_ : int ): logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually snake_case__ : Optional[Any] = self._infer_socket_ifname() # avoid clash with the NCCL port snake_case__ : int = str(distributed_port + 1 ) snake_case__ : List[str] = dist.new_group(ranks=snake_case_ , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def lowerCamelCase ( self : Optional[Any] ): return dist.get_rank(group=self.process_group ) == 0 def lowerCamelCase ( self : int , snake_case_ : str , snake_case_ : int , snake_case_ : int=torch.floataa ): snake_case__ : str = torch.empty(snake_case_ , dtype=snake_case_ ) dist.scatter(snake_case_ , src=0 , scatter_list=snake_case_ , group=self.process_group ) return target_tensor def lowerCamelCase ( self : Union[str, Any] ): snake_case__ : Dict = psutil.net_if_addrs() # a hacky way to deal with varying network interface names snake_case__ : Dict = next((addr for addr in addrs if addr.startswith("""e""" )) , snake_case_ ) return ifname def lowerCamelCase ( self : Tuple , snake_case_ : np.ndarray , snake_case_ : int ): # single GPU training if not dist.is_initialized(): snake_case__ , snake_case__ : Union[str, Any] = self._main_retrieve(snake_case_ , snake_case_ ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(snake_case_ ) # distributed training snake_case__ : Optional[int] = dist.get_world_size(group=self.process_group ) # gather logic snake_case__ : str = None if self._is_main(): snake_case__ : Any = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(snake_case_ )] dist.gather(torch.tensor(snake_case_ ) , dst=0 , gather_list=snake_case_ , group=self.process_group ) # scatter logic snake_case__ : Union[str, Any] = question_hidden_states.shape[0] snake_case__ : List[str] = [] snake_case__ : Dict = [] if self._is_main(): assert len(snake_case_ ) == world_size snake_case__ , snake_case__ : Union[str, Any] = self._main_retrieve(torch.cat(snake_case_ ).numpy() , snake_case_ ) snake_case__ , snake_case__ : Dict = torch.tensor(snake_case_ ), torch.tensor(snake_case_ ) snake_case__ : Union[str, Any] = self._chunk_tensor(snake_case_ , snake_case_ ) snake_case__ : Union[str, Any] = self._chunk_tensor(snake_case_ , snake_case_ ) snake_case__ : Union[str, Any] = self._scattered(snake_case_ , [n_queries, n_docs] , target_type=torch.intaa ) snake_case__ : Dict = self._scattered(snake_case_ , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(snake_case_ )
43
1
from __future__ import annotations import pandas as pd def A_ ( A__ , A__ , A__ ) -> list[int]: a__ : Optional[Any] = [0] * no_of_processes a__ : Tuple = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(A__ ): a__ : Optional[Any] = burst_time[i] a__ : Optional[int] = 0 a__ : int = 0 a__ : int = 9_9999_9999 a__ : Optional[int] = 0 a__ : Tuple = False # Process until all processes are completed while complete != no_of_processes: for j in range(A__ ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: a__ : Dict = remaining_time[j] a__ : int = j a__ : Tuple = True if not check: increment_time += 1 continue remaining_time[short] -= 1 a__ : Tuple = remaining_time[short] if minm == 0: a__ : Optional[Any] = 9_9999_9999 if remaining_time[short] == 0: complete += 1 a__ : Optional[int] = False # Find finish time of current process a__ : Optional[Any] = increment_time + 1 # Calculate waiting time a__ : Optional[Any] = finish_time - arrival_time[short] a__ : Any = finar - burst_time[short] if waiting_time[short] < 0: a__ : int = 0 # Increment time increment_time += 1 return waiting_time def A_ ( A__ , A__ , A__ ) -> list[int]: a__ : List[str] = [0] * no_of_processes for i in range(A__ ): a__ : Dict = burst_time[i] + waiting_time[i] return turn_around_time def A_ ( A__ , A__ , A__ ) -> None: a__ : List[Any] = 0 a__ : Any = 0 for i in range(A__ ): a__ : str = total_waiting_time + waiting_time[i] a__ : Tuple = total_turn_around_time + turn_around_time[i] print(F'Average waiting time = {total_waiting_time / no_of_processes:.5f}' ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") lowercase : str = int(input()) lowercase : List[str] = [0] * no_of_processes lowercase : int = [0] * no_of_processes lowercase : List[Any] = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) lowercase , lowercase : Optional[int] = map(int, input().split()) lowercase : Optional[int] = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowercase : Dict = burst_time lowercase : int = no_of_processes lowercase : Any = waiting_time lowercase : Union[str, Any] = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) lowercase : Any = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
99
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
99
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE :Any = { '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE :List[Any] = [ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE :Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
156
'''simple docstring''' from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING __SCREAMING_SNAKE_CASE :List[Any] = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase_ ) class A_ ( lowerCAmelCase_ ): def __init__( self : List[str] , *snake_case_ : Dict , **snake_case_ : Dict ): super().__init__(*snake_case_ , **snake_case_ ) requires_backends(self , "vision" ) self.check_model_type(snake_case_ ) def __call__( self : Optional[Any] , snake_case_ : Union[str, List[str], "Image.Image", List["Image.Image"]] , **snake_case_ : Optional[int] ): return super().__call__(snake_case_ , **snake_case_ ) def lowercase ( self : Union[str, Any] , **snake_case_ : Union[str, Any] ): return {}, {}, {} def lowercase ( self : Dict , snake_case_ : Optional[int] ): _UpperCAmelCase = load_image(snake_case_ ) _UpperCAmelCase = image.size _UpperCAmelCase = self.image_processor(images=snake_case_ , return_tensors=self.framework ) return model_inputs def lowercase ( self : Optional[int] , snake_case_ : List[Any] ): _UpperCAmelCase = self.model(**snake_case_ ) return model_outputs def lowercase ( self : List[str] , snake_case_ : Dict ): _UpperCAmelCase = model_outputs.predicted_depth _UpperCAmelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode="bicubic" , align_corners=snake_case_ ) _UpperCAmelCase = prediction.squeeze().cpu().numpy() _UpperCAmelCase = (output * 2_5_5 / np.max(snake_case_ )).astype("uint8" ) _UpperCAmelCase = Image.fromarray(snake_case_ ) _UpperCAmelCase = {} _UpperCAmelCase = predicted_depth _UpperCAmelCase = depth return output_dict
156
1
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging UpperCAmelCase : Dict = { "cola": 2, "mnli": 3, "mrpc": 2, "sst-2": 2, "sts-b": 1, "qqp": 2, "qnli": 2, "rte": 2, "wnli": 2, } logging.set_verbosity_info() def __lowerCamelCase ( lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Dict=None ): '''simple docstring''' lowerCamelCase = XLNetConfig.from_json_file(lowerCamelCase__ ) lowerCamelCase = finetuning_task.lower() if finetuning_task is not None else """""" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) lowerCamelCase = finetuning_task lowerCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] lowerCamelCase = XLNetForSequenceClassification(lowerCamelCase__ ) elif "squad" in finetuning_task: lowerCamelCase = finetuning_task lowerCamelCase = XLNetForQuestionAnswering(lowerCamelCase__ ) else: lowerCamelCase = XLNetLMHeadModel(lowerCamelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # Save pytorch-model lowerCamelCase = os.path.join(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase = os.path.join(lowerCamelCase__ , lowerCamelCase__ ) print(f'Save PyTorch model to {os.path.abspath(lowerCamelCase__ )}' ) torch.save(model.state_dict() , lowerCamelCase__ ) print(f'Save configuration file to {os.path.abspath(lowerCamelCase__ )}' ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--xlnet_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained XLNet model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the folder to store the PyTorch model or dataset/vocab.", ) parser.add_argument( "--finetuning_task", default=None, type=str, help="Name of a task on which the XLNet TensorFlow model was fine-tuned", ) UpperCAmelCase : Optional[Any] = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
252
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class __lowercase ( unittest.TestCase ): """simple docstring""" def __A ( self ) -> int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = 1 lowerCamelCase = 3 lowerCamelCase = (32, 32) lowerCamelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(A ) return image @property def __A ( self ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) return model @property def __A ( self ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def __A ( self ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(A ) @property def __A ( self ) -> List[Any]: '''simple docstring''' def extract(*A , **A ): class __lowercase : """simple docstring""" def __init__( self ) -> List[str]: '''simple docstring''' lowerCamelCase = torch.ones([0] ) def __A ( self , A ) -> Tuple: '''simple docstring''' self.pixel_values.to(A ) return self return Out() return extract def __A ( self ) -> Dict: '''simple docstring''' lowerCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.dummy_cond_unet lowerCamelCase = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=A , set_alpha_to_one=A , ) lowerCamelCase = self.dummy_vae lowerCamelCase = self.dummy_text_encoder lowerCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk lowerCamelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = """A painting of a squirrel eating a burger""" lowerCamelCase = torch.Generator(device=A ).manual_seed(0 ) lowerCamelCase = sd_pipe([prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) lowerCamelCase = output.images lowerCamelCase = torch.Generator(device=A ).manual_seed(0 ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=A , )[0] lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ) -> Dict: '''simple docstring''' lowerCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.dummy_cond_unet lowerCamelCase = PNDMScheduler(skip_prk_steps=A ) lowerCamelCase = self.dummy_vae lowerCamelCase = self.dummy_text_encoder lowerCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk lowerCamelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = """A painting of a squirrel eating a burger""" lowerCamelCase = torch.Generator(device=A ).manual_seed(0 ) lowerCamelCase = sd_pipe([prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) lowerCamelCase = output.images lowerCamelCase = torch.Generator(device=A ).manual_seed(0 ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=A , )[0] lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ) -> Tuple: '''simple docstring''' lowerCamelCase = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" , safety_checker=A ) assert isinstance(A , A ) assert isinstance(pipe.scheduler , A ) assert pipe.safety_checker is None lowerCamelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A ) lowerCamelCase = StableDiffusionPipeline.from_pretrained(A ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowerCamelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = self.dummy_cond_unet lowerCamelCase = PNDMScheduler(skip_prk_steps=A ) lowerCamelCase = self.dummy_vae lowerCamelCase = self.dummy_text_encoder lowerCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 lowerCamelCase = unet.half() lowerCamelCase = vae.half() lowerCamelCase = bert.half() # make sure here that pndm scheduler skips prk lowerCamelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = """A painting of a squirrel eating a burger""" lowerCamelCase = sd_pipe([prompt] , num_inference_steps=2 , output_type="""np""" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class __lowercase ( unittest.TestCase ): """simple docstring""" def __A ( self ) -> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=A ) lowerCamelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = ( """portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle""" """ coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with""" """ anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and""" """ children from bahnhof zoo, detailed """ ) lowerCamelCase = 40_03_66_03_46 lowerCamelCase = 7 # without safety guidance (sld_guidance_scale = 0) lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=A ) lowerCamelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = """padme amidala taking a bath artwork, safe for work, no nudity""" lowerCamelCase = 27_34_97_17_55 lowerCamelCase = 7 lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ) -> int: '''simple docstring''' lowerCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) lowerCamelCase = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) lowerCamelCase = ( """the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.""" """ leyendecker""" ) lowerCamelCase = 10_44_35_52_34 lowerCamelCase = 12 lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 lowerCamelCase = torch.manual_seed(A ) lowerCamelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type="""np""" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCamelCase = output.images lowerCamelCase = image[0, -3:, -3:, -1] lowerCamelCase = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
252
1
"""simple docstring""" # Algorithm for the pigeonhole sorting def snake_case__ ( __lowerCamelCase : Tuple ): """simple docstring""" lowerCamelCase__ : Optional[int] =min(__lowerCamelCase ) # min() finds the minimum value lowerCamelCase__ : Tuple =max(__lowerCamelCase ) # max() finds the maximum value lowerCamelCase__ : List[Any] =max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowerCamelCase__ : str =[0] * size # Populate the pigeonholes. for x in a: assert isinstance(__lowerCamelCase , __lowerCamelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowerCamelCase__ : List[str] =0 for count in range(__lowerCamelCase ): while holes[count] > 0: holes[count] -= 1 lowerCamelCase__ : int =count + min_val i += 1 def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : Dict =[8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(__lowerCamelCase ) print('''Sorted order is:''' , ''' '''.join(__lowerCamelCase ) ) if __name__ == "__main__": main()
272
"""simple docstring""" import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def snake_case ( self : Union[str, Any] )-> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() def snake_case ( self : str )-> Any: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] =FlaxControlNetModel.from_pretrained( '''lllyasviel/sd-controlnet-canny''', from_pt=lowerCamelCase, dtype=jnp.bfloataa ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] =FlaxStableDiffusionControlNetPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''', controlnet=lowerCamelCase, from_pt=lowerCamelCase, dtype=jnp.bfloataa ) lowerCamelCase__ : Optional[int] =controlnet_params lowerCamelCase__ : Dict ='''bird''' lowerCamelCase__ : List[str] =jax.device_count() lowerCamelCase__ : Optional[Any] =pipe.prepare_text_inputs([prompts] * num_samples ) lowerCamelCase__ : Dict =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png''' ) lowerCamelCase__ : List[Any] =pipe.prepare_image_inputs([canny_image] * num_samples ) lowerCamelCase__ : Optional[int] =jax.random.PRNGKey(0 ) lowerCamelCase__ : Dict =jax.random.split(lowerCamelCase, jax.device_count() ) lowerCamelCase__ : Tuple =replicate(lowerCamelCase ) lowerCamelCase__ : Tuple =shard(lowerCamelCase ) lowerCamelCase__ : Optional[int] =shard(lowerCamelCase ) lowerCamelCase__ : Tuple =pipe( prompt_ids=lowerCamelCase, image=lowerCamelCase, params=lowerCamelCase, prng_seed=lowerCamelCase, num_inference_steps=50, jit=lowerCamelCase, ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) lowerCamelCase__ : Dict =images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCamelCase__ : Any =images[0, 253:256, 253:256, -1] lowerCamelCase__ : Optional[Any] =jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCamelCase__ : Dict =jnp.array( [0.167_969, 0.116_699, 0.081_543, 0.154_297, 0.132_812, 0.108_887, 0.169_922, 0.169_922, 0.205_078] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def snake_case ( self : Optional[int] )-> Optional[int]: lowerCamelCase__ , lowerCamelCase__ : Dict =FlaxControlNetModel.from_pretrained( '''lllyasviel/sd-controlnet-openpose''', from_pt=lowerCamelCase, dtype=jnp.bfloataa ) lowerCamelCase__ , lowerCamelCase__ : List[Any] =FlaxStableDiffusionControlNetPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''', controlnet=lowerCamelCase, from_pt=lowerCamelCase, dtype=jnp.bfloataa ) lowerCamelCase__ : Optional[Any] =controlnet_params lowerCamelCase__ : int ='''Chef in the kitchen''' lowerCamelCase__ : Optional[Any] =jax.device_count() lowerCamelCase__ : Any =pipe.prepare_text_inputs([prompts] * num_samples ) lowerCamelCase__ : Any =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png''' ) lowerCamelCase__ : List[Any] =pipe.prepare_image_inputs([pose_image] * num_samples ) lowerCamelCase__ : Tuple =jax.random.PRNGKey(0 ) lowerCamelCase__ : Optional[Any] =jax.random.split(lowerCamelCase, jax.device_count() ) lowerCamelCase__ : int =replicate(lowerCamelCase ) lowerCamelCase__ : List[Any] =shard(lowerCamelCase ) lowerCamelCase__ : int =shard(lowerCamelCase ) lowerCamelCase__ : Tuple =pipe( prompt_ids=lowerCamelCase, image=lowerCamelCase, params=lowerCamelCase, prng_seed=lowerCamelCase, num_inference_steps=50, jit=lowerCamelCase, ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) lowerCamelCase__ : Dict =images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCamelCase__ : List[str] =images[0, 253:256, 253:256, -1] lowerCamelCase__ : int =jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCamelCase__ : Any =jnp.array( [[0.271_484, 0.261_719, 0.275_391, 0.277_344, 0.279_297, 0.291_016, 0.294_922, 0.302_734, 0.302_734]] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
272
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: UpperCAmelCase : str = None UpperCAmelCase : Dict = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = {'vocab_file': 'sentencepiece.model', 'tokenizer_file': 'tokenizer.json'} UpperCAmelCase : List[str] = { 'vocab_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/sentencepiece.model', }, 'tokenizer_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/tokenizer.json', }, } UpperCAmelCase : Optional[int] = { 'google/rembert': 2_5_6, } UpperCAmelCase : Any = '▁' class lowerCAmelCase__ ( a ): """simple docstring""" lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = RemBertTokenizer def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : int=None , __SCREAMING_SNAKE_CASE : int=True , __SCREAMING_SNAKE_CASE : List[Any]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : List[str]="[CLS]" , __SCREAMING_SNAKE_CASE : Union[str, Any]="[SEP]" , __SCREAMING_SNAKE_CASE : Union[str, Any]="<unk>" , __SCREAMING_SNAKE_CASE : List[str]="[SEP]" , __SCREAMING_SNAKE_CASE : Union[str, Any]="<pad>" , __SCREAMING_SNAKE_CASE : Any="[CLS]" , __SCREAMING_SNAKE_CASE : Union[str, Any]="[MASK]" , **__SCREAMING_SNAKE_CASE : Union[str, Any] , ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = AddedToken(__SCREAMING_SNAKE_CASE , lstrip=__SCREAMING_SNAKE_CASE , rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else mask_token super().__init__( __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) __SCREAMING_SNAKE_CASE = do_lower_case __SCREAMING_SNAKE_CASE = remove_space __SCREAMING_SNAKE_CASE = keep_accents __SCREAMING_SNAKE_CASE = vocab_file __SCREAMING_SNAKE_CASE = False if not self.vocab_file else True def UpperCAmelCase__ ( self : List[str] , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None ) -> List[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = [self.sep_token_id] __SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None , __SCREAMING_SNAKE_CASE : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( """You should not supply a second sequence if the provided sequence of """ """ids is already formatted with special tokens for the model.""" ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(__SCREAMING_SNAKE_CASE )) + [1] return [1] + ([0] * len(__SCREAMING_SNAKE_CASE )) + [1] def UpperCAmelCase__ ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None ) -> List[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = [self.sep_token_id] __SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self : Dict , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error("""Vocabulary path ({}) should be a directory""".format(__SCREAMING_SNAKE_CASE ) ) return __SCREAMING_SNAKE_CASE = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
267
'''simple docstring''' import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch UpperCAmelCase : int = random.Random() def a__ ( a__ , a__=1.0 , a__=None , a__=None ): """simple docstring""" if rng is None: __SCREAMING_SNAKE_CASE = global_rng __SCREAMING_SNAKE_CASE = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self : str , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : str=7 , __SCREAMING_SNAKE_CASE : List[str]=400 , __SCREAMING_SNAKE_CASE : Any=2_000 , __SCREAMING_SNAKE_CASE : List[str]=10 , __SCREAMING_SNAKE_CASE : Optional[int]=160 , __SCREAMING_SNAKE_CASE : List[str]=8 , __SCREAMING_SNAKE_CASE : Optional[Any]=0.0 , __SCREAMING_SNAKE_CASE : Dict=4_000 , __SCREAMING_SNAKE_CASE : Optional[int]=False , __SCREAMING_SNAKE_CASE : List[Any]=True , ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = min_seq_length __SCREAMING_SNAKE_CASE = max_seq_length __SCREAMING_SNAKE_CASE = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __SCREAMING_SNAKE_CASE = padding_value __SCREAMING_SNAKE_CASE = sampling_rate __SCREAMING_SNAKE_CASE = return_attention_mask __SCREAMING_SNAKE_CASE = do_normalize __SCREAMING_SNAKE_CASE = feature_size __SCREAMING_SNAKE_CASE = chunk_length __SCREAMING_SNAKE_CASE = hop_length def UpperCAmelCase__ ( self : Dict ) -> Dict: """simple docstring""" return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCAmelCase__ ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Union[str, Any]=False , __SCREAMING_SNAKE_CASE : Optional[Any]=False ) -> Union[str, Any]: """simple docstring""" def _flatten(__SCREAMING_SNAKE_CASE : Dict ): return list(itertools.chain(*__SCREAMING_SNAKE_CASE ) ) if equal_length: __SCREAMING_SNAKE_CASE = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __SCREAMING_SNAKE_CASE = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __SCREAMING_SNAKE_CASE = [np.asarray(__SCREAMING_SNAKE_CASE ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCAmelCase__ ( a , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = WhisperFeatureExtractor if is_speech_available() else None def UpperCAmelCase__ ( self : str ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = WhisperFeatureExtractionTester(self ) def UpperCAmelCase__ ( self : str ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __SCREAMING_SNAKE_CASE = feat_extract_first.save_pretrained(__SCREAMING_SNAKE_CASE )[0] check_json_file_has_correct_format(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.feature_extraction_class.from_pretrained(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = feat_extract_first.to_dict() __SCREAMING_SNAKE_CASE = feat_extract_second.to_dict() __SCREAMING_SNAKE_CASE = feat_extract_first.mel_filters __SCREAMING_SNAKE_CASE = feat_extract_second.mel_filters self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : List[str] ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __SCREAMING_SNAKE_CASE = os.path.join(__SCREAMING_SNAKE_CASE , """feat_extract.json""" ) feat_extract_first.to_json_file(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.feature_extraction_class.from_json_file(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = feat_extract_first.to_dict() __SCREAMING_SNAKE_CASE = feat_extract_second.to_dict() __SCREAMING_SNAKE_CASE = feat_extract_first.mel_filters __SCREAMING_SNAKE_CASE = feat_extract_second.mel_filters self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __SCREAMING_SNAKE_CASE = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __SCREAMING_SNAKE_CASE = [np.asarray(__SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs] # Test feature size __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __SCREAMING_SNAKE_CASE = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __SCREAMING_SNAKE_CASE = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test batched __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __SCREAMING_SNAKE_CASE = [floats_list((1, x) )[0] for x in (800, 800, 800)] __SCREAMING_SNAKE_CASE = np.asarray(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test truncation required __SCREAMING_SNAKE_CASE = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __SCREAMING_SNAKE_CASE = [np.asarray(__SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs] __SCREAMING_SNAKE_CASE = [x[: feature_extractor.n_samples] for x in speech_inputs] __SCREAMING_SNAKE_CASE = [np.asarray(__SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs_truncated] __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) ) def UpperCAmelCase__ ( self : Dict ) -> Optional[int]: """simple docstring""" import torch __SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE = np.random.rand(100 , 32 ).astype(np.floataa ) __SCREAMING_SNAKE_CASE = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __SCREAMING_SNAKE_CASE = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __SCREAMING_SNAKE_CASE = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCAmelCase__ ( self : str , __SCREAMING_SNAKE_CASE : Tuple ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __SCREAMING_SNAKE_CASE = ds.sort("""id""" ).select(range(__SCREAMING_SNAKE_CASE ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def UpperCAmelCase__ ( self : Tuple ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on __SCREAMING_SNAKE_CASE = self._load_datasamples(1 ) __SCREAMING_SNAKE_CASE = WhisperFeatureExtractor() __SCREAMING_SNAKE_CASE = feature_extractor(__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3_000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , __SCREAMING_SNAKE_CASE , atol=1E-4 ) ) def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE = self._load_datasamples(1 )[0] __SCREAMING_SNAKE_CASE = ((audio - audio.min()) / (audio.max() - audio.min())) * 65_535 # Rescale to [0, 65535] to show issue __SCREAMING_SNAKE_CASE = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=__SCREAMING_SNAKE_CASE )[0] self.assertTrue(np.all(np.mean(__SCREAMING_SNAKE_CASE ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__SCREAMING_SNAKE_CASE ) - 1 ) < 1E-3 ) )
267
1
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , ) -> float: """simple docstring""" UpperCAmelCase_ : List[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: UpperCAmelCase_ : Dict = 1 - (matter_density + radiation_density + dark_energy) UpperCAmelCase_ : Optional[int] = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) UpperCAmelCase_ : Any = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _lowerCamelCase = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
353
'''simple docstring''' import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def a__ ( _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int ) -> Any: """simple docstring""" if gpta_config_file == "": UpperCAmelCase_ : List[str] = GPTaConfig() else: UpperCAmelCase_ : int = GPTaConfig.from_json_file(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = GPTaModel(_SCREAMING_SNAKE_CASE ) # Load weights from numpy load_tf_weights_in_gpta(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Save pytorch-model UpperCAmelCase_ : Dict = pytorch_dump_folder_path + "/" + WEIGHTS_NAME UpperCAmelCase_ : int = pytorch_dump_folder_path + "/" + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict() , _SCREAMING_SNAKE_CASE ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(_SCREAMING_SNAKE_CASE , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--gpt2_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--gpt2_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _lowerCamelCase = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
67
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ = { "configuration_electra": ["ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP", "ElectraConfig", "ElectraOnnxConfig"], "tokenization_electra": ["ElectraTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = ["ElectraTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "ElectraForCausalLM", "ElectraForMaskedLM", "ElectraForMultipleChoice", "ElectraForPreTraining", "ElectraForQuestionAnswering", "ElectraForSequenceClassification", "ElectraForTokenClassification", "ElectraModel", "ElectraPreTrainedModel", "load_tf_weights_in_electra", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFElectraForMaskedLM", "TFElectraForMultipleChoice", "TFElectraForPreTraining", "TFElectraForQuestionAnswering", "TFElectraForSequenceClassification", "TFElectraForTokenClassification", "TFElectraModel", "TFElectraPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "FlaxElectraForCausalLM", "FlaxElectraForMaskedLM", "FlaxElectraForMultipleChoice", "FlaxElectraForPreTraining", "FlaxElectraForQuestionAnswering", "FlaxElectraForSequenceClassification", "FlaxElectraForTokenClassification", "FlaxElectraModel", "FlaxElectraPreTrainedModel", ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
46
import unittest import numpy as np def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , ) -> np.ndarray: lowerCAmelCase = np.shape(snake_case__ ) lowerCAmelCase = np.shape(snake_case__ ) lowerCAmelCase = np.shape(snake_case__ ) if shape_a[0] != shape_b[0]: lowerCAmelCase = ( '''Expected the same number of rows for A and B. ''' f"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(snake_case__ ) if shape_b[1] != shape_c[1]: lowerCAmelCase = ( '''Expected the same number of columns for B and C. ''' f"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(snake_case__ ) lowerCAmelCase = pseudo_inv if a_inv is None: try: lowerCAmelCase = np.linalg.inv(snake_case__ ) except np.linalg.LinAlgError: raise ValueError( '''Input matrix A is not invertible. Cannot compute Schur complement.''' ) return mat_c - mat_b.T @ a_inv @ mat_b class lowercase_ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1], [6, 3]] ) lowerCAmelCase = schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.block([[a, b], [b.T, c]] ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) self.assertAlmostEqual(__SCREAMING_SNAKE_CASE , det_a * det_s ) def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
338
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Any = { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json''' ), '''google/realm-orqa-nq-openqa''': '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json''', '''google/realm-orqa-nq-reader''': '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json''', '''google/realm-orqa-wq-openqa''': '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json''', '''google/realm-orqa-wq-reader''': '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json''', # See all REALM models at https://huggingface.co/models?filter=realm } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = 'realm' def __init__( self , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=8 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=2_5_6 , lowerCamelCase__=1_0 , lowerCamelCase__=1e-3 , lowerCamelCase__=5 , lowerCamelCase__=3_2_0 , lowerCamelCase__=1_3_3_5_3_7_1_8 , lowerCamelCase__=5_0_0_0 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) # Common config _lowerCamelCase = vocab_size _lowerCamelCase = max_position_embeddings _lowerCamelCase = hidden_size _lowerCamelCase = retriever_proj_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = num_candidates _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = type_vocab_size _lowerCamelCase = layer_norm_eps # Reader config _lowerCamelCase = span_hidden_size _lowerCamelCase = max_span_width _lowerCamelCase = reader_layer_norm_eps _lowerCamelCase = reader_beam_size _lowerCamelCase = reader_seq_len # Retrieval config _lowerCamelCase = num_block_records _lowerCamelCase = searcher_beam_size
362
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
73
0
"""simple docstring""" import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def A ( snake_case :Tuple ) -> List[str]: __UpperCamelCase = SwinConfig() __UpperCamelCase = swin_name.split('_' ) __UpperCamelCase = name_split[1] __UpperCamelCase = int(name_split[4] ) __UpperCamelCase = int(name_split[3][-1] ) if model_size == "tiny": __UpperCamelCase = 9_6 __UpperCamelCase = (2, 2, 6, 2) __UpperCamelCase = (3, 6, 1_2, 2_4) elif model_size == "small": __UpperCamelCase = 9_6 __UpperCamelCase = (2, 2, 1_8, 2) __UpperCamelCase = (3, 6, 1_2, 2_4) elif model_size == "base": __UpperCamelCase = 1_2_8 __UpperCamelCase = (2, 2, 1_8, 2) __UpperCamelCase = (4, 8, 1_6, 3_2) else: __UpperCamelCase = 1_9_2 __UpperCamelCase = (2, 2, 1_8, 2) __UpperCamelCase = (6, 1_2, 2_4, 4_8) if "in22k" in swin_name: __UpperCamelCase = 2_1_8_4_1 else: __UpperCamelCase = 1_0_0_0 __UpperCamelCase = 'huggingface/label-files' __UpperCamelCase = 'imagenet-1k-id2label.json' __UpperCamelCase = json.load(open(hf_hub_download(snake_case , snake_case , repo_type='dataset' ) , 'r' ) ) __UpperCamelCase = {int(snake_case ): v for k, v in idalabel.items()} __UpperCamelCase = idalabel __UpperCamelCase = {v: k for k, v in idalabel.items()} __UpperCamelCase = img_size __UpperCamelCase = num_classes __UpperCamelCase = embed_dim __UpperCamelCase = depths __UpperCamelCase = num_heads __UpperCamelCase = window_size return config def A ( snake_case :Dict ) -> Tuple: if "patch_embed.proj" in name: __UpperCamelCase = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: __UpperCamelCase = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: __UpperCamelCase = 'encoder.' + name if "attn.proj" in name: __UpperCamelCase = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: __UpperCamelCase = name.replace('attn' , 'attention.self' ) if "norm1" in name: __UpperCamelCase = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: __UpperCamelCase = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: __UpperCamelCase = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: __UpperCamelCase = name.replace('mlp.fc2' , 'output.dense' ) if name == "norm.weight": __UpperCamelCase = 'layernorm.weight' if name == "norm.bias": __UpperCamelCase = 'layernorm.bias' if "head" in name: __UpperCamelCase = name.replace('head' , 'classifier' ) else: __UpperCamelCase = 'swin.' + name return name def A ( snake_case :str , snake_case :Any ) -> Any: for key in orig_state_dict.copy().keys(): __UpperCamelCase = orig_state_dict.pop(snake_case ) if "mask" in key: continue elif "qkv" in key: __UpperCamelCase = key.split('.' ) __UpperCamelCase = int(key_split[1] ) __UpperCamelCase = int(key_split[3] ) __UpperCamelCase = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: __UpperCamelCase = val[:dim, :] __UpperCamelCase = val[ dim : dim * 2, : ] __UpperCamelCase = val[-dim:, :] else: __UpperCamelCase = val[ :dim ] __UpperCamelCase = val[ dim : dim * 2 ] __UpperCamelCase = val[ -dim: ] else: __UpperCamelCase = val return orig_state_dict def A ( snake_case :Optional[Any] , snake_case :Optional[int] ) -> Optional[int]: __UpperCamelCase = timm.create_model(snake_case , pretrained=snake_case ) timm_model.eval() __UpperCamelCase = get_swin_config(snake_case ) __UpperCamelCase = SwinForImageClassification(snake_case ) model.eval() __UpperCamelCase = convert_state_dict(timm_model.state_dict() , snake_case ) model.load_state_dict(snake_case ) __UpperCamelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' __UpperCamelCase = AutoImageProcessor.from_pretrained('microsoft/{}'.format(swin_name.replace('_' , '-' ) ) ) __UpperCamelCase = Image.open(requests.get(snake_case , stream=snake_case ).raw ) __UpperCamelCase = image_processor(images=snake_case , return_tensors='pt' ) __UpperCamelCase = timm_model(inputs['pixel_values'] ) __UpperCamelCase = model(**snake_case ).logits assert torch.allclose(snake_case , snake_case , atol=1e-3 ) print(f'Saving model {swin_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case ) if __name__ == "__main__": UpperCamelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swin_name", default="swin_tiny_patch4_window7_224", type=str, help="Name of the Swin timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) UpperCamelCase : Tuple = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
316
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging UpperCamelCase : Any = logging.get_logger(__name__) class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = ["pixel_values"] def __init__( self , __UpperCAmelCase = True , __UpperCAmelCase = 1 / 255 , __UpperCAmelCase = True , __UpperCAmelCase = 8 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) __UpperCamelCase = do_rescale __UpperCamelCase = rescale_factor __UpperCamelCase = do_pad __UpperCamelCase = pad_size def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase ): '''simple docstring''' return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase = get_image_size(__UpperCAmelCase ) __UpperCamelCase = (old_height // size + 1) * size - old_height __UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__UpperCAmelCase , ((0, pad_height), (0, pad_width)) , mode='symmetric' , data_format=__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = ChannelDimension.FIRST , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase = do_pad if do_pad is not None else self.do_pad __UpperCamelCase = pad_size if pad_size is not None else self.pad_size __UpperCamelCase = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) # All transformations expect numpy arrays. __UpperCamelCase = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_rescale: __UpperCamelCase = [self.rescale(image=__UpperCAmelCase , scale=__UpperCAmelCase ) for image in images] if do_pad: __UpperCamelCase = [self.pad(__UpperCAmelCase , size=__UpperCAmelCase ) for image in images] __UpperCamelCase = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] __UpperCamelCase = {'pixel_values': images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
316
1
'''simple docstring''' import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print('Googling.....') SCREAMING_SNAKE_CASE__ = 'https://www.google.com/search?q=' + ' '.join(sys.argv[1:]) SCREAMING_SNAKE_CASE__ = requests.get(url, headers={'UserAgent': UserAgent().random}) # res.raise_for_status() with open('project1a.html', 'wb') as out_file: # only for knowing the class for data in res.iter_content(1_0_0_0_0): out_file.write(data) SCREAMING_SNAKE_CASE__ = BeautifulSoup(res.text, 'html.parser') SCREAMING_SNAKE_CASE__ = list(soup.select('.eZt8xd'))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get('href')) else: webbrowser.open(f'https://google.com{link.get("href")}')
183
'''simple docstring''' import re import string import numpy as np import datasets SCREAMING_SNAKE_CASE__ = '\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n' SCREAMING_SNAKE_CASE__ = '\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results["exact_match"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."]\n >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 33.3\n\n' SCREAMING_SNAKE_CASE__ = '\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def A__ ( self ) -> Optional[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , reference_urls=[] , ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , ) -> List[Any]: """simple docstring""" if regexes_to_ignore is not None: for s in regexes_to_ignore: UpperCamelCase = np.array([re.sub(_SCREAMING_SNAKE_CASE , """""" , _SCREAMING_SNAKE_CASE ) for x in predictions] ) UpperCamelCase = np.array([re.sub(_SCREAMING_SNAKE_CASE , """""" , _SCREAMING_SNAKE_CASE ) for x in references] ) else: UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE ) if ignore_case: UpperCamelCase = np.char.lower(_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.char.lower(_SCREAMING_SNAKE_CASE ) if ignore_punctuation: UpperCamelCase = string.punctuation.maketrans("""""" , """""" , string.punctuation ) UpperCamelCase = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) if ignore_numbers: UpperCamelCase = string.digits.maketrans("""""" , """""" , string.digits ) UpperCamelCase = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) UpperCamelCase = predictions == references return {"exact_match": np.mean(_SCREAMING_SNAKE_CASE ) * 100}
183
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/config.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/config.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/config.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/config.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json", "roberta-large-openai-detector": "https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json", } class lowercase ( a_ ): _a = "roberta" def __init__( self , _a=5_0265 , _a=768 , _a=12 , _a=12 , _a=3072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1e-12 , _a=1 , _a=0 , _a=2 , _a="absolute" , _a=True , _a=None , **_a , ) -> List[str]: super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) _A : Any = vocab_size _A : Optional[int] = hidden_size _A : List[str] = num_hidden_layers _A : List[Any] = num_attention_heads _A : Union[str, Any] = hidden_act _A : Optional[int] = intermediate_size _A : List[Any] = hidden_dropout_prob _A : Any = attention_probs_dropout_prob _A : Tuple = max_position_embeddings _A : Optional[int] = type_vocab_size _A : Any = initializer_range _A : Dict = layer_norm_eps _A : List[str] = position_embedding_type _A : Optional[int] = use_cache _A : Optional[int] = classifier_dropout class lowercase ( a_ ): @property def a__ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _A : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _A : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
26
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("--user", type=str, default="ubuntu") parser.add_argument("--host", type=str, default="localhost") parser.add_argument("--key_path", type=str, default=None) parser.add_argument("--instance", type=str, default="V100:1") parser.add_argument("--provider", type=str, default="cheapest") parser.add_argument("--use_spot", type=bool, default=False) parser.add_argument("--example", type=str, default="pytorch/text-generation/run_generation.py") UpperCAmelCase, UpperCAmelCase : Optional[Any] = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("Cannot specify both BYO and on-demand cluster args") UpperCAmelCase : Dict = rh.cluster( name="rh-cluster", ips=[args.host], ssh_creds={"ssh_user": args.user, "ssh_private_key": args.key_path} ) else: UpperCAmelCase : str = rh.cluster( name="rh-cluster", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) UpperCAmelCase : str = args.example.rsplit("/", 1)[0] # Set up remote environment cluster.install_packages(["pip:./"]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f"""pip install -r transformers/examples/{example_dir}/requirements.txt"""]) cluster.run(["pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f"""python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}"""]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
252
0
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowercase : Any = logging.get_logger(__name__) class A__ ( __UpperCAmelCase ): """simple docstring""" def __init__( self , *lowercase , **lowercase) -> None: '''simple docstring''' warnings.warn( 'The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PoolFormerImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase)
370
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def A_ ( A__ ) -> str: a__ : Any = 384 if "tiny" in model_name: a__ : List[Any] = [3, 3, 9, 3] a__ : Optional[Any] = [96, 192, 384, 768] if "small" in model_name: a__ : Union[str, Any] = [3, 3, 27, 3] a__ : List[Any] = [96, 192, 384, 768] if "base" in model_name: a__ : int = [3, 3, 27, 3] a__ : List[str] = [128, 256, 512, 1024] a__ : Optional[int] = 512 if "large" in model_name: a__ : Optional[int] = [3, 3, 27, 3] a__ : Any = [192, 384, 768, 1536] a__ : int = 768 if "xlarge" in model_name: a__ : str = [3, 3, 27, 3] a__ : int = [256, 512, 1024, 2048] a__ : List[str] = 1024 # set label information a__ : int = 150 a__ : List[Any] = 'huggingface/label-files' a__ : str = 'ade20k-id2label.json' a__ : Optional[int] = json.load(open(hf_hub_download(A__ , A__ , repo_type='dataset' ) , 'r' ) ) a__ : List[str] = {int(A__ ): v for k, v in idalabel.items()} a__ : Union[str, Any] = {v: k for k, v in idalabel.items()} a__ : List[Any] = ConvNextConfig( depths=A__ , hidden_sizes=A__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) a__ : Optional[int] = UperNetConfig( backbone_config=A__ , auxiliary_in_channels=A__ , num_labels=A__ , idalabel=A__ , labelaid=A__ , ) return config def A_ ( A__ ) -> Tuple: a__ : Optional[int] = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.{j}.gamma', F'backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.weight', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.bias', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.weight', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.bias', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias') ) if i > 0: rename_keys.append((F'backbone.downsample_layers.{i}.0.weight', F'backbone.encoder.stages.{i}.downsampling_layer.0.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.0.bias', F'backbone.encoder.stages.{i}.downsampling_layer.0.bias') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.weight', F'backbone.encoder.stages.{i}.downsampling_layer.1.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.bias', F'backbone.encoder.stages.{i}.downsampling_layer.1.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def A_ ( A__ , A__ , A__ ) -> str: a__ : List[str] = dct.pop(A__ ) a__ : int = val def A_ ( A__ , A__ , A__ ) -> str: a__ : Tuple = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } a__ : Dict = model_name_to_url[model_name] a__ : Optional[int] = torch.hub.load_state_dict_from_url(A__ , map_location='cpu' )['state_dict'] a__ : List[Any] = get_upernet_config(A__ ) a__ : Dict = UperNetForSemanticSegmentation(A__ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): a__ : Dict = state_dict.pop(A__ ) if "bn" in key: a__ : Optional[int] = key.replace('bn' , 'batch_norm' ) a__ : List[Any] = val # rename keys a__ : Union[str, Any] = create_rename_keys(A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) model.load_state_dict(A__ ) # verify on image a__ : str = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' a__ : int = Image.open(requests.get(A__ , stream=A__ ).raw ).convert('RGB' ) a__ : Union[str, Any] = SegformerImageProcessor() a__ : Union[str, Any] = processor(A__ , return_tensors='pt' ).pixel_values with torch.no_grad(): a__ : Optional[Any] = model(A__ ) if model_name == "upernet-convnext-tiny": a__ : Union[str, Any] = torch.tensor( [[-8.81_10, -8.81_10, -8.65_21], [-8.81_10, -8.81_10, -8.65_21], [-8.77_46, -8.77_46, -8.61_30]] ) elif model_name == "upernet-convnext-small": a__ : int = torch.tensor( [[-8.82_36, -8.82_36, -8.67_71], [-8.82_36, -8.82_36, -8.67_71], [-8.76_38, -8.76_38, -8.62_40]] ) elif model_name == "upernet-convnext-base": a__ : int = torch.tensor( [[-8.85_58, -8.85_58, -8.69_05], [-8.85_58, -8.85_58, -8.69_05], [-8.76_69, -8.76_69, -8.60_21]] ) elif model_name == "upernet-convnext-large": a__ : Optional[Any] = torch.tensor( [[-8.66_60, -8.66_60, -8.62_10], [-8.66_60, -8.66_60, -8.62_10], [-8.63_10, -8.63_10, -8.59_64]] ) elif model_name == "upernet-convnext-xlarge": a__ : Optional[int] = torch.tensor( [[-8.49_80, -8.49_80, -8.39_77], [-8.49_80, -8.49_80, -8.39_77], [-8.43_79, -8.43_79, -8.34_12]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , A__ , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(A__ ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": lowercase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""upernet-convnext-tiny""", type=str, choices=[F"""upernet-convnext-{size}""" for size in ["""tiny""", """small""", """base""", """large""", """xlarge"""]], help="""Name of the ConvNext UperNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowercase : str = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
225
0
"""simple docstring""" import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline A: Tuple = datasets.utils.logging.get_logger(__name__) @dataclass class SCREAMING_SNAKE_CASE__ ( datasets.BuilderConfig ): __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = 'utf-8' __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Any = True # deprecated __lowerCAmelCase : Optional[Any] = None # deprecated __lowerCAmelCase : Dict = 10 << 20 # 10MB __lowerCAmelCase : Any = None class SCREAMING_SNAKE_CASE__ ( datasets.ArrowBasedBuilder ): __lowerCAmelCase : Union[str, Any] = JsonConfig def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' if self.config.block_size is not None: logger.warning("""The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead""" ) UpperCAmelCase : Dict = self.config.block_size if self.config.use_threads is not True: logger.warning( """The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.""" ) if self.config.newlines_in_values is not None: raise ValueError("""The JSON loader parameter `newlines_in_values` is no longer supported""" ) return datasets.DatasetInfo(features=self.config.features ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F"At least one data file must be specified, but got data_files={self.config.data_files}" ) UpperCAmelCase : Any = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowercase_ , (str, list, tuple) ): UpperCAmelCase : Optional[Any] = data_files if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : str = [files] UpperCAmelCase : Optional[Any] = [dl_manager.iter_files(lowercase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] UpperCAmelCase : List[str] = [] for split_name, files in data_files.items(): if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : List[Any] = [files] UpperCAmelCase : Any = [dl_manager.iter_files(lowercase_ ) for file in files] splits.append(datasets.SplitGenerator(name=lowercase_ , gen_kwargs={"""files""": files} ) ) return splits def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> pa.Table: '''simple docstring''' if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): UpperCAmelCase : Optional[int] = self.config.features.arrow_schema.field(lowercase_ ).type UpperCAmelCase : Optional[Any] = pa_table.append_column(lowercase_ , pa.array([None] * len(lowercase_ ) , type=lowercase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example UpperCAmelCase : int = table_cast(lowercase_ , self.config.features.arrow_schema ) return pa_table def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' for file_idx, file in enumerate(itertools.chain.from_iterable(lowercase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(lowercase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCAmelCase : int = json.load(lowercase_ ) # We keep only the field we are interested in UpperCAmelCase : Tuple = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(lowercase_ , (list, tuple) ): UpperCAmelCase : List[Any] = set().union(*[row.keys() for row in dataset] ) UpperCAmelCase : List[str] = {col: [row.get(lowercase_ ) for row in dataset] for col in keys} else: UpperCAmelCase : Optional[int] = dataset UpperCAmelCase : int = pa.Table.from_pydict(lowercase_ ) yield file_idx, self._cast_table(lowercase_ ) # If the file has one json object per line else: with open(lowercase_ , """rb""" ) as f: UpperCAmelCase : Optional[Any] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small UpperCAmelCase : List[Any] = max(self.config.chunksize // 32 , 16 << 10 ) UpperCAmelCase : Any = ( self.config.encoding_errors if self.config.encoding_errors is not None else """strict""" ) while True: UpperCAmelCase : Any = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(lowercase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": UpperCAmelCase : Union[str, Any] = batch.decode(self.config.encoding , errors=lowercase_ ).encode("""utf-8""" ) try: while True: try: UpperCAmelCase : Optional[int] = paj.read_json( io.BytesIO(lowercase_ ) , read_options=paj.ReadOptions(block_size=lowercase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(lowercase_ , pa.ArrowInvalid ) and "straddling" not in str(lowercase_ ) or block_size > len(lowercase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"Batch of {len(lowercase_ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}." ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( lowercase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCAmelCase : str = json.load(lowercase_ ) except json.JSONDecodeError: logger.error(F"Failed to read file '{file}' with error {type(lowercase_ )}: {e}" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(lowercase_ , lowercase_ ): # list is the only sequence type supported in JSON try: UpperCAmelCase : Optional[int] = set().union(*[row.keys() for row in dataset] ) UpperCAmelCase : List[Any] = {col: [row.get(lowercase_ ) for row in dataset] for col in keys} UpperCAmelCase : str = pa.Table.from_pydict(lowercase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"Failed to read file '{file}' with error {type(lowercase_ )}: {e}" ) raise ValueError(F"Not able to read records in the JSON file at {file}." ) from None yield file_idx, self._cast_table(lowercase_ ) break else: logger.error(F"Failed to read file '{file}' with error {type(lowercase_ )}: {e}" ) raise ValueError( F"Not able to read records in the JSON file at {file}. " F"You should probably indicate the field of the JSON file containing your records. " F"This JSON file contain the following fields: {str(list(dataset.keys() ) )}. " F"Select the correct one and provide it as `field='XXX'` to the dataset loading method. " ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowercase_ ) batch_idx += 1
109
"""simple docstring""" from __future__ import annotations def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase = list(range(len(lowercase_ ) ) ) UpperCAmelCase = [v / w for v, w in zip(lowercase_ , lowercase_ )] index.sort(key=lambda lowercase_ : ratio[i] , reverse=lowercase_ ) UpperCAmelCase = 0 UpperCAmelCase = [0] * len(lowercase_ ) for i in index: if weight[i] <= capacity: UpperCAmelCase = 1 max_value += value[i] capacity -= weight[i] else: UpperCAmelCase = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
78
0
"""simple docstring""" import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class A_ (lowercase__ ): '''simple docstring''' def __init__( self ): """simple docstring""" UpperCAmelCase_ : Dict = [] def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_init_end" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_train_begin" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_train_end" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_epoch_begin" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_epoch_end" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_step_begin" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_step_end" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_evaluate" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_predict" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_save" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_log" ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" self.events.append("on_prediction_step" ) @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = tempfile.mkdtemp() def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.output_dir ) def UpperCamelCase__ ( self , lowercase_=0 , lowercase_=0 , lowercase_=64 , lowercase_=64 , lowercase_=None , lowercase_=False , **lowercase_ ): """simple docstring""" # disable_tqdm in TrainingArguments has a flaky default since it depends on the level of logging. We make sure # its set to False since the tests later on depend on its value. UpperCAmelCase_ : Dict = RegressionDataset(length=lowercase_ ) UpperCAmelCase_ : List[str] = RegressionDataset(length=lowercase_ ) UpperCAmelCase_ : Optional[int] = RegressionModelConfig(a=lowercase_ , b=lowercase_ ) UpperCAmelCase_ : Optional[Any] = RegressionPreTrainedModel(lowercase_ ) UpperCAmelCase_ : Dict = TrainingArguments(self.output_dir , disable_tqdm=lowercase_ , report_to=[] , **lowercase_ ) return Trainer( lowercase_ , lowercase_ , train_dataset=lowercase_ , eval_dataset=lowercase_ , callbacks=lowercase_ , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) # Order doesn't matter UpperCAmelCase_ : Optional[int] = sorted(lowercase_ , key=lambda lowercase_ : cb.__name__ if isinstance(lowercase_ , lowercase_ ) else cb.__class__.__name__ ) UpperCAmelCase_ : Union[str, Any] = sorted(lowercase_ , key=lambda lowercase_ : cb.__name__ if isinstance(lowercase_ , lowercase_ ) else cb.__class__.__name__ ) for cba, cba in zip(lowercase_ , lowercase_ ): if isinstance(lowercase_ , lowercase_ ) and isinstance(lowercase_ , lowercase_ ): self.assertEqual(lowercase_ , lowercase_ ) elif isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): self.assertEqual(lowercase_ , cba.__class__ ) elif not isinstance(lowercase_ , lowercase_ ) and isinstance(lowercase_ , lowercase_ ): self.assertEqual(cba.__class__ , lowercase_ ) else: self.assertEqual(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = ["on_init_end", "on_train_begin"] UpperCAmelCase_ : Union[str, Any] = 0 UpperCAmelCase_ : str = len(trainer.get_eval_dataloader() ) UpperCAmelCase_ : str = ["on_prediction_step"] * len(trainer.get_eval_dataloader() ) + ["on_log", "on_evaluate"] for _ in range(trainer.state.num_train_epochs ): expected_events.append("on_epoch_begin" ) for _ in range(lowercase_ ): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append("on_log" ) if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append("on_save" ) expected_events.append("on_epoch_end" ) if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = self.get_trainer() UpperCAmelCase_ : Union[str, Any] = DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # Callbacks passed at init are added to the default callbacks UpperCAmelCase_ : int = self.get_trainer(callbacks=[MyTestTrainerCallback] ) expected_callbacks.append(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback UpperCAmelCase_ : Any = self.get_trainer(disable_tqdm=lowercase_ ) UpperCAmelCase_ : Tuple = DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = DEFAULT_CALLBACKS.copy() + [ProgressCallback] UpperCAmelCase_ : List[Any] = self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(lowercase_ ) expected_callbacks.remove(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) UpperCAmelCase_ : Dict = self.get_trainer() UpperCAmelCase_ : List[Any] = trainer.pop_callback(lowercase_ ) self.assertEqual(cb.__class__ , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) trainer.add_callback(lowercase_ ) expected_callbacks.insert(0 , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # We can also add, pop, or remove by instance UpperCAmelCase_ : List[Any] = self.get_trainer() UpperCAmelCase_ : Optional[Any] = trainer.callback_handler.callbacks[0] trainer.remove_callback(lowercase_ ) expected_callbacks.remove(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) UpperCAmelCase_ : Dict = self.get_trainer() UpperCAmelCase_ : str = trainer.callback_handler.callbacks[0] UpperCAmelCase_ : List[str] = trainer.pop_callback(lowercase_ ) self.assertEqual(lowercase_ , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) trainer.add_callback(lowercase_ ) expected_callbacks.insert(0 , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action="ignore" , category=lowercase_ ) UpperCAmelCase_ : Optional[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] ) trainer.train() UpperCAmelCase_ : str = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # Independent log/save/eval UpperCAmelCase_ : str = self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5 ) trainer.train() UpperCAmelCase_ : Union[str, Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) UpperCAmelCase_ : Any = self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5 ) trainer.train() UpperCAmelCase_ : Union[str, Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) UpperCAmelCase_ : Optional[int] = self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy="steps" ) trainer.train() UpperCAmelCase_ : List[Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) UpperCAmelCase_ : List[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy="epoch" ) trainer.train() UpperCAmelCase_ : Tuple = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # A bit of everything UpperCAmelCase_ : int = self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=10 , eval_steps=5 , evaluation_strategy="steps" , ) trainer.train() UpperCAmelCase_ : Dict = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # warning should be emitted for duplicated callbacks with patch("transformers.trainer_callback.logger.warning" ) as warn_mock: UpperCAmelCase_ : int = self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(lowercase_ ) in warn_mock.call_args[0][0]
23
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """detr""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=100 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = backbone_config.get("model_type" ) UpperCAmelCase_ : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : Any = config_class.from_dict(lowercase_ ) # set timm attributes to None UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = None, None, None UpperCAmelCase_ : int = use_timm_backbone UpperCAmelCase_ : int = backbone_config UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : int = num_queries UpperCAmelCase_ : Union[str, Any] = d_model UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : List[Any] = encoder_attention_heads UpperCAmelCase_ : Union[str, Any] = decoder_ffn_dim UpperCAmelCase_ : Optional[Any] = decoder_layers UpperCAmelCase_ : Union[str, Any] = decoder_attention_heads UpperCAmelCase_ : Optional[int] = dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Optional[Any] = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Optional[int] = decoder_layerdrop UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : int = auxiliary_loss UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Tuple = backbone UpperCAmelCase_ : Optional[int] = use_pretrained_backbone UpperCAmelCase_ : Dict = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : int = giou_cost # Loss coefficients UpperCAmelCase_ : str = mask_loss_coefficient UpperCAmelCase_ : Any = dice_loss_coefficient UpperCAmelCase_ : Optional[Any] = bbox_loss_coefficient UpperCAmelCase_ : List[str] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): """simple docstring""" return self.d_model @classmethod def UpperCamelCase__ ( cls , lowercase_ , **lowercase_ ): """simple docstring""" return cls(backbone_config=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase_ : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase_ : str = self.__class__.model_type return output class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-5 @property def UpperCamelCase__ ( self ): """simple docstring""" return 12
23
1
def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [0] * len(snake_case_ ) _lowerCAmelCase : Any = [] _lowerCAmelCase : Optional[Any] = [] _lowerCAmelCase : List[str] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(snake_case_ ) ): if indegree[i] == 0: queue.append(snake_case_ ) while queue: _lowerCAmelCase : int = queue.pop(0 ) cnt += 1 topo.append(snake_case_ ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(snake_case_ ) if cnt != len(snake_case_ ): print("Cycle exists" ) else: print(snake_case_ ) # Adjacency List of Graph _snake_case = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
36
from math import loga def lowerCamelCase__ ( snake_case_ : int ) -> int: if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(snake_case_ , snake_case_ ): raise TypeError('''Input value must be a \'int\' type''' ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
24
0
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class UpperCAmelCase ( __A ): '''simple docstring''' def lowerCAmelCase_ ( self ): """simple docstring""" A_ : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_A , 'tf_padding' ) ) self.parent.assertTrue(hasattr(_A , 'depth_multiplier' ) ) class UpperCAmelCase : '''simple docstring''' def __init__( self , lowercase , lowercase=1_3 , lowercase=3 , lowercase=3_2 , lowercase=0.25 , lowercase=8 , lowercase=True , lowercase=1_0_2_4 , lowercase=3_2 , lowercase="relu6" , lowercase=0.1 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=1_0 , lowercase=None , ): """simple docstring""" A_ : int = parent A_ : str = batch_size A_ : Dict = num_channels A_ : Optional[Any] = image_size A_ : Any = depth_multiplier A_ : Any = min_depth A_ : Dict = tf_padding A_ : List[str] = int(last_hidden_size * depth_multiplier ) A_ : Tuple = output_stride A_ : Optional[Any] = hidden_act A_ : Any = classifier_dropout_prob A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = is_training A_ : Optional[int] = num_labels A_ : Tuple = initializer_range A_ : List[Any] = scope def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Optional[Any] = None A_ : Optional[int] = None if self.use_labels: A_ : Dict = ids_tensor([self.batch_size] , self.num_labels ) A_ : int = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A_ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCAmelCase_ ( self ): """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase , lowercase ): """simple docstring""" A_ : int = MobileNetVaModel(config=_A ) model.to(_A ) model.eval() A_ : List[Any] = model(_A ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase , lowercase ): """simple docstring""" A_ : Optional[Any] = self.num_labels A_ : int = MobileNetVaForImageClassification(_A ) model.to(_A ) model.eval() A_ : Any = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Any = self.prepare_config_and_inputs() A_ : Dict = config_and_inputs A_ : Any = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( __A , __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () lowerCamelCase_ = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Optional[int] = MobileNetVaModelTester(self ) A_ : str = MobileNetVaConfigTester(self , config_class=_A , has_text_modality=_A ) def lowerCAmelCase_ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV1 does not use inputs_embeds' ) def lowerCAmelCase_ ( self ): """simple docstring""" pass @unittest.skip(reason='MobileNetV1 does not support input and output embeddings' ) def lowerCAmelCase_ ( self ): """simple docstring""" pass @unittest.skip(reason='MobileNetV1 does not output attentions' ) def lowerCAmelCase_ ( self ): """simple docstring""" pass def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : int = model_class(_A ) A_ : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Tuple = [*signature.parameters.keys()] A_ : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def lowerCAmelCase_ ( self ): """simple docstring""" def check_hidden_states_output(lowercase , lowercase , lowercase ): A_ : Any = model_class(_A ) model.to(_A ) model.eval() with torch.no_grad(): A_ : Dict = model(**self._prepare_for_class(_A , _A ) ) A_ : Dict = outputs.hidden_states A_ : str = 2_6 self.assertEqual(len(_A ) , _A ) A_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = True check_hidden_states_output(_A , _A , _A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_A , _A , _A ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) @slow def lowerCAmelCase_ ( self ): """simple docstring""" for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Tuple = MobileNetVaModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def UpperCamelCase ( ): '''simple docstring''' A_ : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self ): """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v1_1.0_224' ) if is_vision_available() else None ) @slow def lowerCAmelCase_ ( self ): """simple docstring""" A_ : str = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v1_1.0_224' ).to(_A ) A_ : Optional[Any] = self.default_image_processor A_ : str = prepare_img() A_ : Optional[int] = image_processor(images=_A , return_tensors='pt' ).to(_A ) # forward pass with torch.no_grad(): A_ : Union[str, Any] = model(**_A ) # verify the logits A_ : Optional[int] = torch.Size((1, 1_0_0_1) ) self.assertEqual(outputs.logits.shape , _A ) A_ : List[Any] = torch.tensor([-4.1739, -1.1233, 3.1205] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) )
357
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } _UpperCAmelCase = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def UpperCamelCase ( __lowercase : Optional[Any] ,__lowercase : Tuple ,__lowercase : str ,__lowercase : List[Any] ,__lowercase : Optional[int] ): '''simple docstring''' for attribute in key.split('.' ): A_ : List[Any] = getattr(__lowercase ,__lowercase ) if weight_type is not None: A_ : Dict = getattr(__lowercase ,__lowercase ).shape else: A_ : Union[str, Any] = hf_pointer.shape assert hf_shape == value.shape, ( f'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": A_ : Dict = value elif weight_type == "weight_g": A_ : str = value elif weight_type == "weight_v": A_ : int = value elif weight_type == "bias": A_ : int = value else: A_ : List[Any] = value logger.info(f'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def UpperCamelCase ( __lowercase : Optional[Any] ,__lowercase : Union[str, Any] ): '''simple docstring''' A_ : List[Any] = [] A_ : int = fairseq_model.state_dict() A_ : Optional[Any] = hf_model.feature_extractor A_ : List[str] = hf_model.adapter for name, value in fairseq_dict.items(): A_ : Tuple = False if "conv_layers" in name: load_conv_layer( __lowercase ,__lowercase ,__lowercase ,__lowercase ,hf_model.config.feat_extract_norm == 'group' ,) A_ : Optional[Any] = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(__lowercase ,__lowercase ,__lowercase ,__lowercase ) A_ : Union[str, Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: A_ : Tuple = True if "*" in mapped_key: A_ : Optional[Any] = name.split(__lowercase )[0].split('.' )[-2] A_ : List[Any] = mapped_key.replace('*' ,__lowercase ) if "weight_g" in name: A_ : Optional[int] = 'weight_g' elif "weight_v" in name: A_ : Union[str, Any] = 'weight_v' elif "bias" in name: A_ : Any = 'bias' elif "weight" in name: A_ : str = 'weight' else: A_ : Optional[Any] = None set_recursively(__lowercase ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) continue if not is_used: unused_weights.append(__lowercase ) logger.warning(f'''Unused weights: {unused_weights}''' ) def UpperCamelCase ( __lowercase : Dict ,__lowercase : List[Any] ,__lowercase : Tuple ,__lowercase : Dict ,__lowercase : Any ): '''simple docstring''' A_ : List[Any] = full_name.split('conv_layers.' )[-1] A_ : Optional[int] = name.split('.' ) A_ : Tuple = int(items[0] ) A_ : int = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) A_ : Union[str, Any] = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) A_ : str = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) A_ : List[str] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) A_ : List[Any] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__lowercase ) def UpperCamelCase ( __lowercase : Any ,__lowercase : Tuple ,__lowercase : Optional[Any] ,__lowercase : Union[str, Any] ): '''simple docstring''' A_ : Union[str, Any] = full_name.split('adaptor.' )[-1] A_ : List[Any] = name.split('.' ) if items[1].isdigit(): A_ : Union[str, Any] = int(items[1] ) else: A_ : Tuple = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.''' A_ : Dict = value logger.info(f'''Adapter proj layer norm bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.''' A_ : int = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.''' A_ : Dict = value logger.info(f'''Adapter proj layer bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.''' A_ : Tuple = value logger.info(f'''Adapter proj layer weight was initialized from {full_name}.''' ) elif isinstance(__lowercase ,__lowercase ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.''' A_ : Tuple = value logger.info(f'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.''' A_ : str = value logger.info(f'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) else: unused_weights.append(__lowercase ) def UpperCamelCase ( __lowercase : List[Any] ): '''simple docstring''' A_ , A_ : Any = emb.weight.shape A_ : Tuple = nn.Linear(__lowercase ,__lowercase ,bias=__lowercase ) A_ : Optional[Any] = emb.weight.data return lin_layer @torch.no_grad() def UpperCamelCase ( __lowercase : Any ,__lowercase : Optional[int] ,__lowercase : Any ,__lowercase : str ,__lowercase : Dict ,__lowercase : Dict ,__lowercase : Tuple ,__lowercase : Optional[int] ,__lowercase : List[str] ,__lowercase : List[Any] ,__lowercase : str ,): '''simple docstring''' A_ : Optional[int] = WavaVecaConfig.from_pretrained( __lowercase ,add_adapter=__lowercase ,adapter_stride=__lowercase ,adapter_kernel_size=__lowercase ,use_auth_token=__lowercase ,output_hidden_size=__lowercase ,) A_ : Any = MBartConfig.from_pretrained(__lowercase ) # load model A_ , A_ , A_ : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } ,) A_ : Union[str, Any] = model[0].eval() # load feature extractor A_ : Any = WavaVecaFeatureExtractor.from_pretrained(__lowercase ,use_auth_token=__lowercase ) # set weights for wav2vec2 encoder A_ : Optional[Any] = WavaVecaModel(__lowercase ) recursively_load_weights_wavaveca(model.encoder ,__lowercase ) # load decoder weights A_ : Dict = MBartForCausalLM(__lowercase ) A_ , A_ : Union[str, Any] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() ,strict=__lowercase ) logger.warning(f'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(f'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) A_ : Optional[int] = SpeechEncoderDecoderModel(encoder=__lowercase ,decoder=__lowercase ) A_ : Any = False A_ : List[Any] = MBartaaTokenizer(__lowercase ) tokenizer.save_pretrained(__lowercase ) A_ : Dict = hf_wavavec.config.to_dict() A_ : Any = tokenizer.pad_token_id A_ : Optional[Any] = tokenizer.bos_token_id A_ : Union[str, Any] = tokenizer.eos_token_id A_ : Dict = 'mbart50' A_ : str = 'wav2vec2' A_ : int = tokenizer.eos_token_id A_ : List[str] = 25_00_04 A_ : int = tokenizer.eos_token_id A_ : Optional[Any] = SpeechEncoderDecoderConfig.from_dict(__lowercase ) hf_wavavec.save_pretrained(__lowercase ) feature_extractor.save_pretrained(__lowercase ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_yaml_path""", default=None, type=str, help="""Path to yaml file of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-xls-r-1b""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/mbart-large-50-one-to-many-mmt""", type=str, help="""Path to hf decoder checkpoint config""", ) parser.add_argument("""--add_adapter""", default=True, type=bool, help="""whethere to add model adapter layers""") parser.add_argument("""--adapter_stride""", default=2, type=int, help="""stride of adapter layers""") parser.add_argument("""--adapter_kernel_size""", default=3, type=int, help="""kernel size of adapter layers""") parser.add_argument("""--encoder_output_dim""", default=1024, type=int, help="""encoder output dim""") parser.add_argument("""--start_token_id""", default=250004, type=int, help="""`decoder_start_token_id` of model config""") _UpperCAmelCase = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
192
0
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class __UpperCamelCase : """simple docstring""" def __init__( self : Optional[int] , _A : int , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Tuple = parent __SCREAMING_SNAKE_CASE : Any = 13 __SCREAMING_SNAKE_CASE : Tuple = 7 __SCREAMING_SNAKE_CASE : Tuple = True __SCREAMING_SNAKE_CASE : Dict = True __SCREAMING_SNAKE_CASE : int = True __SCREAMING_SNAKE_CASE : str = 99 __SCREAMING_SNAKE_CASE : Dict = 32 __SCREAMING_SNAKE_CASE : List[str] = 2 __SCREAMING_SNAKE_CASE : int = 4 __SCREAMING_SNAKE_CASE : List[Any] = 37 __SCREAMING_SNAKE_CASE : Optional[int] = '''gelu''' __SCREAMING_SNAKE_CASE : Dict = 0.1 __SCREAMING_SNAKE_CASE : Any = 0.1 __SCREAMING_SNAKE_CASE : str = 512 __SCREAMING_SNAKE_CASE : Optional[int] = 16 __SCREAMING_SNAKE_CASE : Optional[int] = 2 __SCREAMING_SNAKE_CASE : List[Any] = 0.02 __SCREAMING_SNAKE_CASE : int = 3 __SCREAMING_SNAKE_CASE : List[str] = 4 __SCREAMING_SNAKE_CASE : List[str] = None def UpperCAmelCase__ ( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE : List[Any] = None if self.use_input_mask: __SCREAMING_SNAKE_CASE : Dict = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : List[str] = None __SCREAMING_SNAKE_CASE : Union[str, Any] = None if self.use_labels: __SCREAMING_SNAKE_CASE : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE : Any = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE : Dict = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" ( ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ) : List[str] = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE : Optional[int] = True __SCREAMING_SNAKE_CASE : Any = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __SCREAMING_SNAKE_CASE : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCAmelCase__ ( self : List[Any] , _A : Optional[Any] , _A : List[Any] , _A : Tuple , _A : Union[str, Any] , _A : List[str] , _A : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = TFEsmModel(config=_A ) __SCREAMING_SNAKE_CASE : Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __SCREAMING_SNAKE_CASE : Tuple = model(_A ) __SCREAMING_SNAKE_CASE : Optional[Any] = [input_ids, input_mask] __SCREAMING_SNAKE_CASE : str = model(_A ) __SCREAMING_SNAKE_CASE : Dict = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase__ ( self : Tuple , _A : Optional[Any] , _A : Dict , _A : int , _A : Dict , _A : str , _A : Optional[Any] , _A : int , _A : int , ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = True __SCREAMING_SNAKE_CASE : Dict = TFEsmModel(config=_A ) __SCREAMING_SNAKE_CASE : Optional[int] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } __SCREAMING_SNAKE_CASE : List[str] = model(_A ) __SCREAMING_SNAKE_CASE : Any = [input_ids, input_mask] __SCREAMING_SNAKE_CASE : Optional[Any] = model(_A , encoder_hidden_states=_A ) # Also check the case where encoder outputs are not passed __SCREAMING_SNAKE_CASE : str = model(_A , attention_mask=_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase__ ( self : List[str] , _A : Union[str, Any] , _A : Optional[Any] , _A : str , _A : Union[str, Any] , _A : str , _A : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = TFEsmForMaskedLM(config=_A ) __SCREAMING_SNAKE_CASE : int = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase__ ( self : Tuple , _A : Optional[Any] , _A : int , _A : Any , _A : List[str] , _A : List[Any] , _A : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = self.num_labels __SCREAMING_SNAKE_CASE : Optional[Any] = TFEsmForTokenClassification(config=_A ) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __SCREAMING_SNAKE_CASE : Tuple = model(_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ) : int = config_and_inputs __SCREAMING_SNAKE_CASE : str = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase_ = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase_ = ( { '''feature-extraction''': TFEsmModel, '''fill-mask''': TFEsmForMaskedLM, '''text-classification''': TFEsmForSequenceClassification, '''token-classification''': TFEsmForTokenClassification, '''zero-shot''': TFEsmForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False def UpperCAmelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = TFEsmModelTester(self ) __SCREAMING_SNAKE_CASE : List[Any] = ConfigTester(self , config_class=_A , hidden_size=37 ) def UpperCAmelCase__ ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_A ) def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_A ) def UpperCAmelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_A ) @slow def UpperCAmelCase__ ( self : Any ): """simple docstring""" for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE : Dict = TFEsmModel.from_pretrained(_A ) self.assertIsNotNone(_A ) @unittest.skip('''Protein models do not support embedding resizing.''' ) def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" pass @unittest.skip('''Protein models do not support embedding resizing.''' ) def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" pass def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE : Optional[int] = model_class(_A ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer __SCREAMING_SNAKE_CASE : int = model.get_bias() assert isinstance(_A , _A ) for k, v in name.items(): assert isinstance(_A , tf.Variable ) else: __SCREAMING_SNAKE_CASE : int = model.get_output_embeddings() assert x is None __SCREAMING_SNAKE_CASE : Optional[Any] = model.get_bias() assert name is None @require_tf class __UpperCamelCase ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = TFEsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __SCREAMING_SNAKE_CASE : int = tf.constant([[0, 1, 2, 3, 4, 5]] ) __SCREAMING_SNAKE_CASE : Optional[int] = model(_A )[0] __SCREAMING_SNAKE_CASE : Tuple = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , _A ) # compare the actual values for a slice. __SCREAMING_SNAKE_CASE : Dict = tf.constant( [ [ [8.92_15_18, -10.58_98_14, -6.4_67_13_07], [-6.3_96_71_56, -13.91_13_77, -1.1_21_19_15], [-7.78_12_47, -13.95_15_57, -3.74_05_92], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-2 ) ) @slow def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = TFEsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __SCREAMING_SNAKE_CASE : Any = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) __SCREAMING_SNAKE_CASE : List[str] = model(_A )[0] # compare the actual values for a slice. __SCREAMING_SNAKE_CASE : Union[str, Any] = tf.constant( [ [ [0.14_44_30_92, 0.54_12_53_27, 0.3_24_77_39], [0.30_34_04_84, 0.00_52_66_76, 0.31_07_77_22], [0.32_27_80_43, -0.24_98_70_96, 0.3_41_46_28], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
303
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem lowercase_ = importlib.util.find_spec("""s3fs""") is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 lowercase_ = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(f'''A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.''') fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a__ ( snake_case ): """simple docstring""" if "://" in dataset_path: __SCREAMING_SNAKE_CASE : Any = dataset_path.split('''://''' )[1] return dataset_path def a__ ( snake_case ): """simple docstring""" if fs is not None and fs.protocol != "file": return True else: return False def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = not is_remote_filesystem(snake_case ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(snake_case ) , fs._strip_protocol(snake_case ) ) else: fs.mv(snake_case , snake_case , recursive=snake_case ) def a__ ( ): """simple docstring""" if hasattr(fsspec.asyn , '''reset_lock''' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __SCREAMING_SNAKE_CASE : int = None __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Union[str, Any] = threading.Lock()
303
1
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : str , lowercase : Tuple=None , lowercase : Optional[int]=None ): '''simple docstring''' if attention_mask is None: lowerCamelCase_ = tf.cast(tf.math.not_equal(lowercase , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class A: '''simple docstring''' UpperCamelCase = OPTConfig UpperCamelCase = {} UpperCamelCase = '''gelu''' def __init__( self : str , A_ : Optional[int] , A_ : Union[str, Any]=13 , A_ : Any=7 , A_ : Any=True , A_ : Optional[int]=False , A_ : int=99 , A_ : Tuple=16 , A_ : Dict=2 , A_ : Dict=4 , A_ : Optional[Any]=4 , A_ : Union[str, Any]="gelu" , A_ : List[Any]=0.1 , A_ : List[str]=0.1 , A_ : str=20 , A_ : Union[str, Any]=2 , A_ : Optional[int]=1 , A_ : List[Any]=0 , A_ : Dict=16 , A_ : Union[str, Any]=16 , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = seq_length lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = vocab_size lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = max_position_embeddings lowerCamelCase_ = eos_token_id lowerCamelCase_ = pad_token_id lowerCamelCase_ = bos_token_id lowerCamelCase_ = embed_dim lowerCamelCase_ = word_embed_proj_dim lowerCamelCase_ = False def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCamelCase_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCamelCase_ = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCamelCase_ = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=A_ , **self.config_updates , ) lowerCamelCase_ = prepare_opt_inputs_dict(A_ , A_ ) return config, inputs_dict def a__ ( self : Optional[int] , A_ : Optional[int] , A_ : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFOPTModel(config=A_ ) lowerCamelCase_ = inputs_dict['input_ids'] lowerCamelCase_ = input_ids[:1, :] lowerCamelCase_ = inputs_dict['attention_mask'][:1, :] lowerCamelCase_ = 1 # first forward pass lowerCamelCase_ = model(A_ , attention_mask=A_ , use_cache=A_ ) lowerCamelCase_ , lowerCamelCase_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCamelCase_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCamelCase_ = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCamelCase_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCamelCase_ = model(A_ , attention_mask=A_ )[0] lowerCamelCase_ = model(A_ , attention_mask=A_ , past_key_values=A_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCamelCase_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCamelCase_ = output_from_no_past[:, -3:, random_slice_idx] lowerCamelCase_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A_ , A_ , rtol=1E-3 ) @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () UpperCamelCase = (TFOPTForCausalLM,) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFOPTModel, '''text-generation''': TFOPTForCausalLM} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = 10 def a__ ( self : str ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = TFOPTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ ) def a__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" self.config_tester.run_common_tests() def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A_ ) def a__ ( self : Dict ) -> Optional[int]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(A_ : Optional[int] , A_ : List[Any] ): if hasattr(A_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(A_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings lowerCamelCase_ = model_class(config=A_ ) lowerCamelCase_ = _get_word_embedding_weight(A_ , model.get_input_embeddings() ) lowerCamelCase_ = _get_word_embedding_weight(A_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(A_ ) lowerCamelCase_ = _get_word_embedding_weight(A_ , model.get_input_embeddings() ) lowerCamelCase_ = _get_word_embedding_weight(A_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. lowerCamelCase_ = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , A_ ) # check that weights remain the same after resizing lowerCamelCase_ = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: lowerCamelCase_ = False self.assertTrue(A_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , A_ ) lowerCamelCase_ = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: lowerCamelCase_ = False self.assertTrue(A_ ) def _SCREAMING_SNAKE_CASE ( lowercase : Optional[int] ): '''simple docstring''' return tf.constant(lowercase , dtype=tf.intaa ) @require_tf class A( unittest.TestCase ): '''simple docstring''' UpperCamelCase = 99 def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = tf.ones((4, 1) , dtype=tf.intaa ) * 2 lowerCamelCase_ = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) lowerCamelCase_ = input_ids.shape[0] lowerCamelCase_ = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class A( unittest.TestCase ): '''simple docstring''' @slow def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = TFOPTModel.from_pretrained('facebook/opt-350m' ) lowerCamelCase_ = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) lowerCamelCase_ = tf.not_equal(A_ , model.config.pad_token_id ) with tf.GradientTape(): lowerCamelCase_ = model(input_ids=A_ , attention_mask=A_ ).last_hidden_state lowerCamelCase_ = (1, 11, 512) self.assertEqual(output.shape , A_ ) lowerCamelCase_ = tf.constant( [[-0.2873, -1.9218, -0.3033], [-1.2710, -0.1338, -0.1902], [0.4095, 0.1214, -1.3121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , A_ , atol=4E-3 ) ) lowerCamelCase_ = tf.function(A_ , jit_compile=A_ ) lowerCamelCase_ = xla_generate(A_ , A_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , A_ , atol=4E-2 ) ) @require_tf @slow class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Any ) -> Dict: """simple docstring""" super().setUp() lowerCamelCase_ = 'facebook/opt-350m' def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = TFOPTForCausalLM.from_pretrained(self.path_model ) lowerCamelCase_ = GPTaTokenizer.from_pretrained(self.path_model ) lowerCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False lowerCamelCase_ = tokenizer(A_ , return_tensors='tf' , padding=A_ , add_special_tokens=A_ ) lowerCamelCase_ = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) lowerCamelCase_ = tf.constant( [ [1.3851, -13.8923, -10.5229, -10.7533, -0.2309, -10.2384, -0.5365, -9.0947, -5.1670], [-4.7073, -10.6276, -3.9415, -21.5242, -0.2822, -0.2822, -0.2822, -0.2822, -0.2822], [0.6247, -3.4229, -8.9179, -1.4297, -14.1650, 1.4146, -9.0218, -0.2703, -0.2703], [6.4783, -1.9913, -10.7926, -2.3336, 1.5092, -0.9974, -6.8213, 1.3477, 1.3477], ] ) self.assertTrue(np.allclose(A_ , A_ , atol=1E-4 ) ) lowerCamelCase_ = tf.function(A_ , jit_compile=A_ ) lowerCamelCase_ = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(A_ , A_ , atol=1E-4 ) ) @require_tf @slow class A( unittest.TestCase ): '''simple docstring''' @property def a__ ( self : List[Any] ) -> Dict: """simple docstring""" return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = 'facebook/opt-125m' lowerCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] lowerCamelCase_ = [] lowerCamelCase_ = GPTaTokenizer.from_pretrained(A_ ) lowerCamelCase_ = TFOPTForCausalLM.from_pretrained(A_ ) for prompt in self.prompts: lowerCamelCase_ = tokenizer(A_ , return_tensors='tf' ).input_ids lowerCamelCase_ = model.generate(A_ , max_length=10 ) lowerCamelCase_ = tokenizer.batch_decode(A_ , skip_special_tokens=A_ ) predicted_outputs += generated_string self.assertListEqual(A_ , A_ ) def a__ ( self : Optional[int] ) -> List[str]: """simple docstring""" lowerCamelCase_ = 'facebook/opt-350m' lowerCamelCase_ = GPTaTokenizer.from_pretrained(A_ ) lowerCamelCase_ = TFOPTForCausalLM.from_pretrained(A_ ) lowerCamelCase_ = 'left' # use different length sentences to test batching lowerCamelCase_ = [ 'Hello, my dog is a little', 'Today, I', ] lowerCamelCase_ = tokenizer(A_ , return_tensors='tf' , padding=A_ ) lowerCamelCase_ = inputs['input_ids'] lowerCamelCase_ = model.generate(input_ids=A_ , attention_mask=inputs['attention_mask'] ) lowerCamelCase_ = tokenizer(sentences[0] , return_tensors='tf' ).input_ids lowerCamelCase_ = model.generate(input_ids=A_ ) lowerCamelCase_ = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) lowerCamelCase_ = tokenizer(sentences[1] , return_tensors='tf' ).input_ids lowerCamelCase_ = model.generate(input_ids=A_ , max_length=model.config.max_length - num_paddings ) lowerCamelCase_ = tokenizer.batch_decode(A_ , skip_special_tokens=A_ ) lowerCamelCase_ = tokenizer.decode(output_non_padded[0] , skip_special_tokens=A_ ) lowerCamelCase_ = tokenizer.decode(output_padded[0] , skip_special_tokens=A_ ) lowerCamelCase_ = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , [non_padded_sentence, padded_sentence] ) def a__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = 'facebook/opt-350m' lowerCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] lowerCamelCase_ = [] lowerCamelCase_ = GPTaTokenizer.from_pretrained(A_ ) lowerCamelCase_ = TFOPTForCausalLM.from_pretrained(A_ ) for prompt in self.prompts: lowerCamelCase_ = tokenizer(A_ , return_tensors='tf' ).input_ids lowerCamelCase_ = model.generate(A_ , max_length=10 ) lowerCamelCase_ = tokenizer.batch_decode(A_ , skip_special_tokens=A_ ) predicted_outputs += generated_string self.assertListEqual(A_ , A_ )
367
import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, get_image_size, infer_channel_dimension_format, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_vision_available, logging from ...utils.import_utils import requires_backends if is_vision_available(): import textwrap from PIL import Image, ImageDraw, ImageFont if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowerCamelCase : Optional[Any] = False lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase : Any = "ybelkada/fonts" def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' if is_torch_available() and not is_torch_greater_or_equal_than_1_11: raise ImportError( f"""You are using torch=={torch.__version__}, but torch>=1.11.0 is required to use """ 'Pix2StructImageProcessor. Please upgrade torch.' ) def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : List[str] , lowercase : List[Any] ): '''simple docstring''' requires_backends(lowercase , ['torch'] ) _check_torch_version() lowerCamelCase_ = image_tensor.unsqueeze(0 ) lowerCamelCase_ = torch.nn.functional.unfold(lowercase , (patch_height, patch_width) , stride=(patch_height, patch_width) ) lowerCamelCase_ = patches.reshape(image_tensor.size(0 ) , image_tensor.size(1 ) , lowercase , lowercase , -1 ) lowerCamelCase_ = patches.permute(0 , 4 , 2 , 3 , 1 ).reshape( image_tensor.size(2 ) // patch_height , image_tensor.size(3 ) // patch_width , image_tensor.size(1 ) * patch_height * patch_width , ) return patches.unsqueeze(0 ) def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int = 36 , lowercase : str = "black" , lowercase : str = "white" , lowercase : int = 5 , lowercase : int = 5 , lowercase : int = 5 , lowercase : int = 5 , lowercase : Optional[bytes] = None , lowercase : Optional[str] = None , ): '''simple docstring''' requires_backends(lowercase , 'vision' ) # Add new lines so that each line is no more than 80 characters. lowerCamelCase_ = textwrap.TextWrapper(width=80 ) lowerCamelCase_ = wrapper.wrap(text=lowercase ) lowerCamelCase_ = '\n'.join(lowercase ) if font_bytes is not None and font_path is None: lowerCamelCase_ = io.BytesIO(lowercase ) elif font_path is not None: lowerCamelCase_ = font_path else: lowerCamelCase_ = hf_hub_download(lowercase , 'Arial.TTF' ) lowerCamelCase_ = ImageFont.truetype(lowercase , encoding='UTF-8' , size=lowercase ) # Use a temporary canvas to determine the width and height in pixels when # rendering the text. lowerCamelCase_ = ImageDraw.Draw(Image.new('RGB' , (1, 1) , lowercase ) ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = temp_draw.textbbox((0, 0) , lowercase , lowercase ) # Create the actual image with a bit of padding around the text. lowerCamelCase_ = text_width + left_padding + right_padding lowerCamelCase_ = text_height + top_padding + bottom_padding lowerCamelCase_ = Image.new('RGB' , (image_width, image_height) , lowercase ) lowerCamelCase_ = ImageDraw.Draw(lowercase ) draw.text(xy=(left_padding, top_padding) , text=lowercase , fill=lowercase , font=lowercase ) return image def _SCREAMING_SNAKE_CASE ( lowercase : np.ndarray , lowercase : str , **lowercase : List[Any] ): '''simple docstring''' requires_backends(lowercase , 'vision' ) # Convert to PIL image if necessary lowerCamelCase_ = to_pil_image(lowercase ) lowerCamelCase_ = render_text(lowercase , **lowercase ) lowerCamelCase_ = max(header_image.width , image.width ) lowerCamelCase_ = int(image.height * (new_width / image.width) ) lowerCamelCase_ = int(header_image.height * (new_width / header_image.width) ) lowerCamelCase_ = Image.new('RGB' , (new_width, new_height + new_header_height) , 'white' ) new_image.paste(header_image.resize((new_width, new_header_height) ) , (0, 0) ) new_image.paste(image.resize((new_width, new_height) ) , (0, new_header_height) ) # Convert back to the original framework if necessary lowerCamelCase_ = to_numpy_array(lowercase ) if infer_channel_dimension_format(lowercase ) == ChannelDimension.LAST: lowerCamelCase_ = to_channel_dimension_format(lowercase , ChannelDimension.LAST ) return new_image class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = ['''flattened_patches'''] def __init__( self : Dict , A_ : bool = True , A_ : bool = True , A_ : Dict[str, int] = None , A_ : int = 2048 , A_ : bool = False , **A_ : str , ) -> None: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = patch_size if patch_size is not None else {'height': 16, 'width': 16} lowerCamelCase_ = do_normalize lowerCamelCase_ = do_convert_rgb lowerCamelCase_ = max_patches lowerCamelCase_ = is_vqa def a__ ( self : Union[str, Any] , A_ : np.ndarray , A_ : int , A_ : dict , **A_ : Any ) -> np.ndarray: """simple docstring""" requires_backends(self.extract_flattened_patches , 'torch' ) _check_torch_version() # convert to torch lowerCamelCase_ = to_channel_dimension_format(A_ , ChannelDimension.FIRST ) lowerCamelCase_ = torch.from_numpy(A_ ) lowerCamelCase_ , lowerCamelCase_ = patch_size['height'], patch_size['width'] lowerCamelCase_ , lowerCamelCase_ = get_image_size(A_ ) # maximize scale s.t. lowerCamelCase_ = math.sqrt(max_patches * (patch_height / image_height) * (patch_width / image_width) ) lowerCamelCase_ = max(min(math.floor(scale * image_height / patch_height ) , A_ ) , 1 ) lowerCamelCase_ = max(min(math.floor(scale * image_width / patch_width ) , A_ ) , 1 ) lowerCamelCase_ = max(num_feasible_rows * patch_height , 1 ) lowerCamelCase_ = max(num_feasible_cols * patch_width , 1 ) lowerCamelCase_ = torch.nn.functional.interpolate( image.unsqueeze(0 ) , size=(resized_height, resized_width) , mode='bilinear' , align_corners=A_ , antialias=A_ , ).squeeze(0 ) # [1, rows, columns, patch_height * patch_width * image_channels] lowerCamelCase_ = torch_extract_patches(A_ , A_ , A_ ) lowerCamelCase_ = patches.shape lowerCamelCase_ = patches_shape[1] lowerCamelCase_ = patches_shape[2] lowerCamelCase_ = patches_shape[3] # [rows * columns, patch_height * patch_width * image_channels] lowerCamelCase_ = patches.reshape([rows * columns, depth] ) # [rows * columns, 1] lowerCamelCase_ = torch.arange(A_ ).reshape([rows, 1] ).repeat(1 , A_ ).reshape([rows * columns, 1] ) lowerCamelCase_ = torch.arange(A_ ).reshape([1, columns] ).repeat(A_ , 1 ).reshape([rows * columns, 1] ) # Offset by 1 so the ids do not contain zeros, which represent padding. row_ids += 1 col_ids += 1 # Prepare additional patch features. # [rows * columns, 1] lowerCamelCase_ = row_ids.to(torch.floataa ) lowerCamelCase_ = col_ids.to(torch.floataa ) # [rows * columns, 2 + patch_height * patch_width * image_channels] lowerCamelCase_ = torch.cat([row_ids, col_ids, patches] , -1 ) # [max_patches, 2 + patch_height * patch_width * image_channels] lowerCamelCase_ = torch.nn.functional.pad(A_ , [0, 0, 0, max_patches - (rows * columns)] ).float() lowerCamelCase_ = to_numpy_array(A_ ) return result def a__ ( self : Optional[Any] , A_ : np.ndarray , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : str ) -> np.ndarray: """simple docstring""" if image.dtype == np.uinta: lowerCamelCase_ = image.astype(np.floataa ) # take mean across the whole `image` lowerCamelCase_ = np.mean(A_ ) lowerCamelCase_ = np.std(A_ ) lowerCamelCase_ = max(A_ , 1.0 / math.sqrt(np.prod(image.shape ) ) ) return normalize(A_ , mean=A_ , std=A_ , **A_ ) def a__ ( self : Optional[Any] , A_ : ImageInput , A_ : Optional[str] = None , A_ : bool = None , A_ : Optional[bool] = None , A_ : Optional[int] = None , A_ : Optional[Dict[str, int]] = None , A_ : Optional[Union[str, TensorType]] = None , A_ : ChannelDimension = ChannelDimension.FIRST , **A_ : Optional[int] , ) -> ImageInput: """simple docstring""" lowerCamelCase_ = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase_ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCamelCase_ = patch_size if patch_size is not None else self.patch_size lowerCamelCase_ = max_patches if max_patches is not None else self.max_patches lowerCamelCase_ = self.is_vqa if kwargs.get('data_format' , A_ ) is not None: raise ValueError('data_format is not an accepted input as the outputs are ' ) lowerCamelCase_ = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCamelCase_ = [convert_to_rgb(A_ ) for image in images] # All transformations expect numpy arrays. lowerCamelCase_ = [to_numpy_array(A_ ) for image in images] if is_vqa: if header_text is None: raise ValueError('A header text must be provided for VQA models.' ) lowerCamelCase_ = kwargs.pop('font_bytes' , A_ ) lowerCamelCase_ = kwargs.pop('font_path' , A_ ) if isinstance(A_ , A_ ): lowerCamelCase_ = [header_text] * len(A_ ) lowerCamelCase_ = [ render_header(A_ , header_text[i] , font_bytes=A_ , font_path=A_ ) for i, image in enumerate(A_ ) ] if do_normalize: lowerCamelCase_ = [self.normalize(image=A_ ) for image in images] # convert to torch tensor and permute lowerCamelCase_ = [ self.extract_flattened_patches(image=A_ , max_patches=A_ , patch_size=A_ ) for image in images ] # create attention mask in numpy lowerCamelCase_ = [(image.sum(axis=-1 ) != 0).astype(np.floataa ) for image in images] lowerCamelCase_ = BatchFeature( data={'flattened_patches': images, 'attention_mask': attention_masks} , tensor_type=A_ ) return encoded_outputs
208
0
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model __lowercase = '''0.12''' # assumed parallelism: 8 if is_torch_available(): import torch def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): '''simple docstring''' if rng is None: __UpperCamelCase :Union[str, Any] = random.Random() __UpperCamelCase :int = 1 for dim in shape: total_dims *= dim __UpperCamelCase :str = [] for _ in range(SCREAMING_SNAKE_CASE ): values.append(rng.randint(0 , vocab_size - 1 ) ) __UpperCamelCase :List[str] = np.array(SCREAMING_SNAKE_CASE , dtype=jnp.intaa ).reshape(SCREAMING_SNAKE_CASE ) return output def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): '''simple docstring''' __UpperCamelCase :Optional[Any] = ids_tensor(SCREAMING_SNAKE_CASE , vocab_size=2 , rng=SCREAMING_SNAKE_CASE ) # make sure that at least one token is attended to for each batch __UpperCamelCase :List[Any] = 1 return attn_mask @require_flax class lowerCamelCase_ : '''simple docstring''' a__ : List[Any] = None a__ : Union[str, Any] = () def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase , __UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 __UpperCamelCase :List[str] = 2 __UpperCamelCase :Tuple = inputs['''input_ids'''].shape[-1] // 2 __UpperCamelCase :Optional[Any] = inputs['''input_ids'''][:max_batch_size, :sequence_length] __UpperCamelCase :str = jnp.ones_like(__lowercase) __UpperCamelCase :int = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens __UpperCamelCase :Dict = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` __UpperCamelCase :List[str] = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def UpperCamelCase__ ( self) -> str: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self._get_input_ids_and_config() __UpperCamelCase :Any = False __UpperCamelCase :Tuple = max_length __UpperCamelCase :Optional[int] = 0 for model_class in self.all_generative_model_classes: __UpperCamelCase :List[Any] = model_class(__lowercase) __UpperCamelCase :Dict = model_class.__name__[4:] # Skip the "Flax" at the beginning __UpperCamelCase :List[str] = getattr(__lowercase , __lowercase) __UpperCamelCase :Tuple = pt_model_class(__lowercase).eval() __UpperCamelCase :Tuple = load_flax_weights_in_pytorch_model(__lowercase , flax_model.params) __UpperCamelCase :int = flax_model.generate(__lowercase).sequences __UpperCamelCase :List[str] = pt_model.generate(torch.tensor(__lowercase , dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: __UpperCamelCase :Any = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = self._get_input_ids_and_config() __UpperCamelCase :Tuple = False __UpperCamelCase :str = max_length for model_class in self.all_generative_model_classes: __UpperCamelCase :Union[str, Any] = model_class(__lowercase) __UpperCamelCase :Optional[Any] = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :Dict = jit(model.generate) __UpperCamelCase :str = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Tuple = self._get_input_ids_and_config() __UpperCamelCase :int = True __UpperCamelCase :List[Any] = max_length for model_class in self.all_generative_model_classes: __UpperCamelCase :List[str] = model_class(__lowercase) __UpperCamelCase :List[str] = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :int = jit(model.generate) __UpperCamelCase :List[Any] = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> int: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[int] = self._get_input_ids_and_config() __UpperCamelCase :Any = False __UpperCamelCase :Any = max_length __UpperCamelCase :Any = 2 for model_class in self.all_generative_model_classes: __UpperCamelCase :Optional[int] = model_class(__lowercase) __UpperCamelCase :Tuple = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :Optional[int] = jit(model.generate) __UpperCamelCase :Dict = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self._get_input_ids_and_config() __UpperCamelCase :Any = False __UpperCamelCase :Any = max_length __UpperCamelCase :int = 2 __UpperCamelCase :List[str] = 2 for model_class in self.all_generative_model_classes: __UpperCamelCase :Any = model_class(__lowercase) __UpperCamelCase :Union[str, Any] = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[Any] = self._get_input_ids_and_config() __UpperCamelCase :Tuple = True __UpperCamelCase :int = max_length __UpperCamelCase :Tuple = 0.8 __UpperCamelCase :Union[str, Any] = 10 __UpperCamelCase :List[Any] = 0.3 __UpperCamelCase :Any = 1 __UpperCamelCase :str = 8 __UpperCamelCase :Dict = 9 for model_class in self.all_generative_model_classes: __UpperCamelCase :str = model_class(__lowercase) __UpperCamelCase :Any = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :Union[str, Any] = jit(model.generate) __UpperCamelCase :Union[str, Any] = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Dict = self._get_input_ids_and_config() __UpperCamelCase :int = max_length __UpperCamelCase :str = 1 __UpperCamelCase :Any = 8 __UpperCamelCase :str = 9 for model_class in self.all_generative_model_classes: __UpperCamelCase :Union[str, Any] = model_class(__lowercase) __UpperCamelCase :str = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :List[Any] = jit(model.generate) __UpperCamelCase :Union[str, Any] = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Dict = self._get_input_ids_and_config() __UpperCamelCase :Union[str, Any] = max_length __UpperCamelCase :Union[str, Any] = 2 __UpperCamelCase :List[str] = 1 __UpperCamelCase :Tuple = 8 __UpperCamelCase :Optional[int] = 9 for model_class in self.all_generative_model_classes: __UpperCamelCase :List[Any] = model_class(__lowercase) __UpperCamelCase :str = model.generate(__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :List[str] = jit(model.generate) __UpperCamelCase :Dict = jit_generate(__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[int] = self._get_input_ids_and_config() # pad attention mask on the left __UpperCamelCase :List[Any] = attention_mask.at[(0, 0)].set(0) __UpperCamelCase :int = False __UpperCamelCase :List[str] = max_length for model_class in self.all_generative_model_classes: __UpperCamelCase :Tuple = model_class(__lowercase) __UpperCamelCase :int = model.generate(__lowercase , attention_mask=__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :Optional[Any] = jit(model.generate) __UpperCamelCase :str = jit_generate(__lowercase , attention_mask=__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[Any] = self._get_input_ids_and_config() # pad attention mask on the left __UpperCamelCase :int = attention_mask.at[(0, 0)].set(0) __UpperCamelCase :int = True __UpperCamelCase :int = max_length for model_class in self.all_generative_model_classes: __UpperCamelCase :Union[str, Any] = model_class(__lowercase) __UpperCamelCase :str = model.generate(__lowercase , attention_mask=__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :Any = jit(model.generate) __UpperCamelCase :int = jit_generate(__lowercase , attention_mask=__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[str] = self._get_input_ids_and_config() # pad attention mask on the left __UpperCamelCase :Dict = attention_mask.at[(0, 0)].set(0) __UpperCamelCase :Optional[Any] = 2 __UpperCamelCase :List[str] = max_length for model_class in self.all_generative_model_classes: __UpperCamelCase :Union[str, Any] = model_class(__lowercase) __UpperCamelCase :List[Any] = model.generate(__lowercase , attention_mask=__lowercase).sequences self.assertEqual(generation_outputs.shape[-1] , __lowercase) __UpperCamelCase :int = jit(model.generate) __UpperCamelCase :int = jit_generate(__lowercase , attention_mask=__lowercase).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) @require_flax class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Optional[int] = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') __UpperCamelCase :Any = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') __UpperCamelCase :Optional[Any] = '''Hello world''' __UpperCamelCase :int = tokenizer(__lowercase , return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(__lowercase , '''do_samples'''): model.generate(__lowercase , do_samples=__lowercase) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(__lowercase , '''foo'''): __UpperCamelCase :List[Any] = {'''foo''': '''bar'''} model.generate(__lowercase , **__lowercase)
43
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
1
"""simple docstring""" import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _A = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. _A = importlib.util.spec_from_file_location( """transformers""", os.path.join(PATH_TO_TRANSFORMERS, """__init__.py"""), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) _A = spec.loader.load_module() _A = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` _A = re.compile("""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") _A = { """CLIPConfigMixin""", """DecisionTransformerConfigMixin""", """EncoderDecoderConfigMixin""", """RagConfigMixin""", """SpeechEncoderDecoderConfigMixin""", """VisionEncoderDecoderConfigMixin""", """VisionTextDualEncoderConfigMixin""", } def lowercase_ ( ) -> Union[str, Any]: lowerCAmelCase__ : int = [] for config_class in list(CONFIG_MAPPING.values() ): lowerCAmelCase__ : Dict = False # source code of `config_class` lowerCAmelCase__ : Dict = inspect.getsource(__UpperCAmelCase ) lowerCAmelCase__ : List[str] = _re_checkpoint.findall(__UpperCAmelCase ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` lowerCAmelCase__ : Tuple = checkpoint # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Union[str, Any] = f"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : List[str] = True break lowerCAmelCase__ : List[str] = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: lowerCAmelCase__ : Dict = """\n""".join(sorted(__UpperCAmelCase ) ) raise ValueError(f"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
367
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _A = logging.get_logger(__name__) class _lowerCamelCase ( a_ ): def __init__( self : Union[str, Any] , *UpperCamelCase : int , **UpperCamelCase : List[Any] ) -> None: """simple docstring""" warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCamelCase , ) super().__init__(*UpperCamelCase , **UpperCamelCase )
212
0