code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule __snake_case = {'''tokenization_wav2vec2_phoneme''': ['''Wav2Vec2PhonemeCTCTokenizer''']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
97
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def a ( __a , __a ) -> Optional[int]: '''simple docstring''' assert isinstance(__a , __a ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def a ( __a , __a , __a ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = tmp_path / '''cache''' UpperCamelCase__ :Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase__ :Tuple = JsonDatasetReader(__a , cache_dir=__a , keep_in_memory=__a ).read() _check_json_dataset(__a , __a ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def a ( __a , __a , __a ) -> Any: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = tmp_path / '''cache''' UpperCamelCase__ :Optional[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCamelCase__ :Optional[Any] = features.copy() if features else default_expected_features UpperCamelCase__ :Tuple = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase__ :int = JsonDatasetReader(__a , features=__a , cache_dir=__a ).read() _check_json_dataset(__a , __a ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def a ( __a , __a , __a ) -> Tuple: '''simple docstring''' UpperCamelCase__ :int = tmp_path / '''cache''' UpperCamelCase__ :str = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCamelCase__ :Any = features.copy() if features else default_expected_features UpperCamelCase__ :Union[str, Any] = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase__ :Any = JsonDatasetReader(__a , features=__a , cache_dir=__a ).read() assert isinstance(__a , __a ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def a ( __a , __a ) -> List[Any]: '''simple docstring''' UpperCamelCase__ :Any = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCamelCase__ :int = features.copy() UpperCamelCase__ :List[Any] = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase__ :Optional[int] = tmp_path / '''cache''' UpperCamelCase__ :Dict = JsonDatasetReader(__a , features=__a , cache_dir=__a ).read() assert isinstance(__a , __a ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def a ( __a , __a , __a ) -> List[Any]: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = tmp_path / '''cache''' UpperCamelCase__ :Optional[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCamelCase__ :List[Any] = JsonDatasetReader(__a , cache_dir=__a , split=__a ).read() _check_json_dataset(__a , __a ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def a ( __a , __a , __a ) -> Any: '''simple docstring''' if issubclass(__a , __a ): UpperCamelCase__ :Union[str, Any] = jsonl_path elif issubclass(__a , __a ): UpperCamelCase__ :int = [jsonl_path] UpperCamelCase__ :Dict = tmp_path / '''cache''' UpperCamelCase__ :Any = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCamelCase__ :List[str] = JsonDatasetReader(__a , cache_dir=__a ).read() _check_json_dataset(__a , __a ) def a ( __a , __a , __a=("train",) ) -> Optional[Any]: '''simple docstring''' assert isinstance(__a , __a ) for split in splits: UpperCamelCase__ :Optional[int] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def a ( __a , __a , __a ) -> List[str]: '''simple docstring''' UpperCamelCase__ :List[str] = tmp_path / '''cache''' UpperCamelCase__ :Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase__ :str = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=__a , keep_in_memory=__a ).read() _check_json_datasetdict(__a , __a ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def a ( __a , __a , __a ) -> int: '''simple docstring''' UpperCamelCase__ :Tuple = tmp_path / '''cache''' UpperCamelCase__ :Any = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCamelCase__ :Optional[int] = features.copy() if features else default_expected_features UpperCamelCase__ :str = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase__ :Dict = JsonDatasetReader({'''train''': jsonl_path} , features=__a , cache_dir=__a ).read() _check_json_datasetdict(__a , __a ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def a ( __a , __a , __a ) -> str: '''simple docstring''' if split: UpperCamelCase__ :List[str] = {split: jsonl_path} else: UpperCamelCase__ :int = '''train''' UpperCamelCase__ :int = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCamelCase__ :Any = tmp_path / '''cache''' UpperCamelCase__ :Union[str, Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCamelCase__ :Any = JsonDatasetReader(__a , cache_dir=__a ).read() _check_json_datasetdict(__a , __a , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def a ( __a ) -> Union[str, Any]: '''simple docstring''' return json.load(__a ) def a ( __a ) -> int: '''simple docstring''' return [json.loads(__a ) for line in buffer] class lowercase : """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , lines=UpperCamelCase_ ).write() buffer.seek(0 ) UpperCamelCase__ :List[Any] = load_json_function(UpperCamelCase_ ) assert isinstance(UpperCamelCase_ , UpperCamelCase_ ) assert isinstance(exported_content[0] , UpperCamelCase_ ) assert len(UpperCamelCase_ ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , lines=UpperCamelCase_ , orient=UpperCamelCase_ ).write() buffer.seek(0 ) UpperCamelCase__ :Optional[int] = load_json(UpperCamelCase_ ) assert isinstance(UpperCamelCase_ , UpperCamelCase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(UpperCamelCase_ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(UpperCamelCase_ ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , lines=UpperCamelCase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCamelCase__ :Union[str, Any] = load_json_function(UpperCamelCase_ ) assert isinstance(UpperCamelCase_ , UpperCamelCase_ ) assert isinstance(exported_content[0] , UpperCamelCase_ ) assert len(UpperCamelCase_ ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , lines=UpperCamelCase_ , orient=UpperCamelCase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCamelCase__ :int = load_json(UpperCamelCase_ ) assert isinstance(UpperCamelCase_ , UpperCamelCase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(UpperCamelCase_ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(UpperCamelCase_ ) == 10 def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' with pytest.raises(UpperCamelCase_ ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Tuple = tmp_path_factory.mktemp('''data''' ) / F'''test.json.{extension}''' UpperCamelCase__ :Union[str, Any] = str(shared_datadir / F'''test_file.json.{extension}''' ) JsonDatasetWriter(UpperCamelCase_ , UpperCamelCase_ , compression=UpperCamelCase_ ).write() with fsspec.open(UpperCamelCase_ , '''rb''' , compression='''infer''' ) as f: UpperCamelCase__ :Dict = f.read() with fsspec.open(UpperCamelCase_ , '''rb''' , compression='''infer''' ) as f: UpperCamelCase__ :int = f.read() assert exported_content == original_content
97
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "xlm-mlm-en-2048": "https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json", "xlm-mlm-ende-1024": "https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json", "xlm-mlm-enfr-1024": "https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json", "xlm-mlm-enro-1024": "https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json", "xlm-mlm-tlm-xnli15-1024": "https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json", "xlm-mlm-xnli15-1024": "https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json", "xlm-clm-enfr-1024": "https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json", "xlm-clm-ende-1024": "https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json", "xlm-mlm-17-1280": "https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json", "xlm-mlm-100-1280": "https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json", } class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :int = "xlm" _UpperCAmelCase :List[Any] = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self , _UpperCAmelCase=30145 , _UpperCAmelCase=2048 , _UpperCAmelCase=12 , _UpperCAmelCase=16 , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase=False , _UpperCAmelCase=False , _UpperCAmelCase=1 , _UpperCAmelCase=True , _UpperCAmelCase=512 , _UpperCAmelCase=2048**-0.5 , _UpperCAmelCase=1e-1_2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=0 , _UpperCAmelCase=1 , _UpperCAmelCase=2 , _UpperCAmelCase=3 , _UpperCAmelCase=5 , _UpperCAmelCase=True , _UpperCAmelCase="first" , _UpperCAmelCase=True , _UpperCAmelCase=None , _UpperCAmelCase=True , _UpperCAmelCase=0.1 , _UpperCAmelCase=5 , _UpperCAmelCase=5 , _UpperCAmelCase=0 , _UpperCAmelCase=0 , _UpperCAmelCase=2 , _UpperCAmelCase=0 , **_UpperCAmelCase , ): lowercase__: Optional[Any] = vocab_size lowercase__: Any = emb_dim lowercase__: Optional[int] = n_layers lowercase__: Optional[int] = n_heads lowercase__: Dict = dropout lowercase__: str = attention_dropout lowercase__: Union[str, Any] = gelu_activation lowercase__: Union[str, Any] = sinusoidal_embeddings lowercase__: Union[str, Any] = causal lowercase__: int = asm lowercase__: Tuple = n_langs lowercase__: Tuple = use_lang_emb lowercase__: Union[str, Any] = layer_norm_eps lowercase__: List[str] = bos_index lowercase__: Optional[Any] = eos_index lowercase__: int = pad_index lowercase__: Union[str, Any] = unk_index lowercase__: Union[str, Any] = mask_index lowercase__: Tuple = is_encoder lowercase__: Dict = max_position_embeddings lowercase__: Optional[int] = embed_init_std lowercase__: Tuple = init_std lowercase__: List[Any] = summary_type lowercase__: List[str] = summary_use_proj lowercase__: Optional[int] = summary_activation lowercase__: Any = summary_proj_to_labels lowercase__: Any = summary_first_dropout lowercase__: List[str] = start_n_top lowercase__: Any = end_n_top lowercase__: Optional[Any] = mask_token_id lowercase__: Union[str, Any] = lang_id if "n_words" in kwargs: lowercase__: Tuple = kwargs['''n_words'''] super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" @property def _snake_case ( self ): if self.task == "multiple-choice": lowercase__: int = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase__: int = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
355
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase : """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=0.2 , _UpperCAmelCase=0.2 ): lowercase__: int = bp_numa lowercase__: Union[str, Any] = bp_numa lowercase__: List[str] = bp_numa lowercase__: str = conva_get[:2] lowercase__: Union[str, Any] = conva_get[2] lowercase__: Any = size_pa lowercase__: Optional[Any] = rate_w lowercase__: Tuple = rate_t lowercase__: List[str] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__: Dict = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__: str = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__: Union[str, Any] = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__: Any = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__: Any = -2 * np.random.rand(self.num_bpa ) + 1 def _snake_case ( self , _UpperCAmelCase ): # save model dict with pickle lowercase__: int = { '''num_bp1''': self.num_bpa, '''num_bp2''': self.num_bpa, '''num_bp3''': self.num_bpa, '''conv1''': self.conva, '''step_conv1''': self.step_conva, '''size_pooling1''': self.size_poolinga, '''rate_weight''': self.rate_weight, '''rate_thre''': self.rate_thre, '''w_conv1''': self.w_conva, '''wkj''': self.wkj, '''vji''': self.vji, '''thre_conv1''': self.thre_conva, '''thre_bp2''': self.thre_bpa, '''thre_bp3''': self.thre_bpa, } with open(_UpperCAmelCase , '''wb''' ) as f: pickle.dump(_UpperCAmelCase , _UpperCAmelCase ) print(F"""Model saved: {save_path}""" ) @classmethod def _snake_case ( cls , _UpperCAmelCase ): # read saved model with open(_UpperCAmelCase , '''rb''' ) as f: lowercase__: Optional[int] = pickle.load(_UpperCAmelCase ) # noqa: S301 lowercase__: Tuple = model_dic.get('''conv1''' ) conv_get.append(model_dic.get('''step_conv1''' ) ) lowercase__: Any = model_dic.get('''size_pooling1''' ) lowercase__: int = model_dic.get('''num_bp1''' ) lowercase__: Optional[int] = model_dic.get('''num_bp2''' ) lowercase__: str = model_dic.get('''num_bp3''' ) lowercase__: Any = model_dic.get('''rate_weight''' ) lowercase__: Union[str, Any] = model_dic.get('''rate_thre''' ) # create model instance lowercase__: str = CNN(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # modify model parameter lowercase__: Dict = model_dic.get('''w_conv1''' ) lowercase__: Dict = model_dic.get('''wkj''' ) lowercase__: str = model_dic.get('''vji''' ) lowercase__: List[Any] = model_dic.get('''thre_conv1''' ) lowercase__: Optional[int] = model_dic.get('''thre_bp2''' ) lowercase__: Tuple = model_dic.get('''thre_bp3''' ) return conv_ins def _snake_case ( self , _UpperCAmelCase ): return 1 / (1 + np.exp(-1 * x )) def _snake_case ( self , _UpperCAmelCase ): return round(_UpperCAmelCase , 3 ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): # convolution process lowercase__: Any = convs[0] lowercase__: Tuple = convs[1] lowercase__: List[Any] = np.shape(_UpperCAmelCase )[0] # get the data slice of original image data, data_focus lowercase__: List[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , _UpperCAmelCase ): for j_focus in range(0 , size_data - size_conv + 1 , _UpperCAmelCase ): lowercase__: Tuple = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(_UpperCAmelCase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__: Optional[int] = [] lowercase__: Optional[int] = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(_UpperCAmelCase ): lowercase__: str = [] for i_focus in range(len(_UpperCAmelCase ) ): lowercase__: Any = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(_UpperCAmelCase ) ) lowercase__: str = np.asmatrix(_UpperCAmelCase ).reshape( _UpperCAmelCase , _UpperCAmelCase ) data_featuremap.append(_UpperCAmelCase ) # expanding the data slice to One dimenssion lowercase__: Union[str, Any] = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(_UpperCAmelCase ) ) lowercase__: Any = np.asarray(_UpperCAmelCase ) return focus_list, data_featuremap def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase="average_pool" ): # pooling process lowercase__: List[Any] = len(featuremaps[0] ) lowercase__: Any = int(size_map / size_pooling ) lowercase__: List[Any] = [] for i_map in range(len(_UpperCAmelCase ) ): lowercase__: Any = featuremaps[i_map] lowercase__: Tuple = [] for i_focus in range(0 , _UpperCAmelCase , _UpperCAmelCase ): for j_focus in range(0 , _UpperCAmelCase , _UpperCAmelCase ): lowercase__: Optional[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(_UpperCAmelCase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(_UpperCAmelCase ) ) lowercase__: str = np.asmatrix(_UpperCAmelCase ).reshape(_UpperCAmelCase , _UpperCAmelCase ) featuremap_pooled.append(_UpperCAmelCase ) return featuremap_pooled def _snake_case ( self , _UpperCAmelCase ): # expanding three dimension data to one dimension list lowercase__: Optional[Any] = [] for i in range(len(_UpperCAmelCase ) ): lowercase__: Any = np.shape(data[i] ) lowercase__: List[Any] = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__: List[str] = data_listed.getA().tolist()[0] data_expanded.extend(_UpperCAmelCase ) lowercase__: List[str] = np.asarray(_UpperCAmelCase ) return data_expanded def _snake_case ( self , _UpperCAmelCase ): # expanding matrix to one dimension list lowercase__: Union[str, Any] = np.asarray(_UpperCAmelCase ) lowercase__: List[str] = np.shape(_UpperCAmelCase ) lowercase__: List[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowercase__: str = [] lowercase__: List[str] = 0 for i_map in range(_UpperCAmelCase ): lowercase__: Union[str, Any] = np.ones((size_map, size_map) ) for i in range(0 , _UpperCAmelCase , _UpperCAmelCase ): for j in range(0 , _UpperCAmelCase , _UpperCAmelCase ): lowercase__: Optional[Any] = pd_pool[ i_pool ] lowercase__: List[Any] = i_pool + 1 lowercase__: str = np.multiply( _UpperCAmelCase , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(_UpperCAmelCase ) return pd_all def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=bool ): # model traning print('''----------------------Start Training-------------------------''' ) print((''' - - Shape: Train_Data ''', np.shape(_UpperCAmelCase )) ) print((''' - - Shape: Teach_Data ''', np.shape(_UpperCAmelCase )) ) lowercase__: Tuple = 0 lowercase__: Tuple = [] lowercase__: Optional[int] = 10000 while rp < n_repeat and mse >= error_accuracy: lowercase__: Tuple = 0 print(F"""-------------Learning Time {rp}--------------""" ) for p in range(len(_UpperCAmelCase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__: List[Any] = np.asmatrix(datas_train[p] ) lowercase__: Optional[int] = np.asarray(datas_teach[p] ) lowercase__, lowercase__: List[str] = self.convolute( _UpperCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__: Optional[int] = self.pooling(_UpperCAmelCase , self.size_poolinga ) lowercase__: int = np.shape(_UpperCAmelCase ) lowercase__: Optional[Any] = self._expand(_UpperCAmelCase ) lowercase__: Any = data_bp_input lowercase__: Any = np.dot(_UpperCAmelCase , self.vji.T ) - self.thre_bpa lowercase__: str = self.sig(_UpperCAmelCase ) lowercase__: Optional[Any] = np.dot(_UpperCAmelCase , self.wkj.T ) - self.thre_bpa lowercase__: Dict = self.sig(_UpperCAmelCase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__: str = np.multiply( (data_teach - bp_outa) , np.multiply(_UpperCAmelCase , (1 - bp_outa) ) ) lowercase__: str = np.multiply( np.dot(_UpperCAmelCase , self.wkj ) , np.multiply(_UpperCAmelCase , (1 - bp_outa) ) ) lowercase__: Dict = np.dot(_UpperCAmelCase , self.vji ) lowercase__: Any = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__: List[str] = pd_conva_pooled.T.getA().tolist() lowercase__: Optional[Any] = self._calculate_gradient_from_pool( _UpperCAmelCase , _UpperCAmelCase , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__: str = self._expand_mat(pd_conva_all[k_conv] ) lowercase__: str = self.rate_weight * np.dot(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Any = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__: List[Any] = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__: Optional[int] = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__: List[Any] = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__: List[str] = self.thre_bpa - pd_k_all * self.rate_thre lowercase__: Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__: Optional[Any] = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__: str = rp + 1 lowercase__: Optional[Any] = error_count / patterns all_mse.append(_UpperCAmelCase ) def draw_error(): lowercase__: Union[str, Any] = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(_UpperCAmelCase , '''+-''' ) plt.plot(_UpperCAmelCase , '''r--''' ) plt.xlabel('''Learning Times''' ) plt.ylabel('''All_mse''' ) plt.grid(_UpperCAmelCase , alpha=0.5 ) plt.show() print('''------------------Training Complished---------------------''' ) print((''' - - Training epoch: ''', rp, F""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _snake_case ( self , _UpperCAmelCase ): # model predict lowercase__: Union[str, Any] = [] print('''-------------------Start Testing-------------------------''' ) print((''' - - Shape: Test_Data ''', np.shape(_UpperCAmelCase )) ) for p in range(len(_UpperCAmelCase ) ): lowercase__: Union[str, Any] = np.asmatrix(datas_test[p] ) lowercase__, lowercase__: Any = self.convolute( _UpperCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__: List[str] = self.pooling(_UpperCAmelCase , self.size_poolinga ) lowercase__: str = self._expand(_UpperCAmelCase ) lowercase__: List[Any] = data_bp_input lowercase__: List[str] = bp_outa * self.vji.T - self.thre_bpa lowercase__: Any = self.sig(_UpperCAmelCase ) lowercase__: Optional[int] = bp_outa * self.wkj.T - self.thre_bpa lowercase__: Any = self.sig(_UpperCAmelCase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__: str = [list(map(self.do_round , _UpperCAmelCase ) ) for each in produce_out] return np.asarray(_UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase ): # return the data of image after convoluting process so we can check it out lowercase__: int = np.asmatrix(_UpperCAmelCase ) lowercase__, lowercase__: Optional[int] = self.convolute( _UpperCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__: List[Any] = self.pooling(_UpperCAmelCase , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
2
0
def a__ ( snake_case , snake_case , snake_case = 0 , snake_case = 0 ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = right or len(snake_case ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(snake_case , snake_case , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
303
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
303
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class lowercase( a_ ): '''simple docstring''' lowercase__ = "realm" def __init__( self: List[str], a_: Optional[Any]=30_522, a_: Any=768, a_: Tuple=128, a_: Optional[int]=12, a_: int=12, a_: List[Any]=8, a_: Optional[Any]=3_072, a_: Any="gelu_new", a_: List[str]=0.1, a_: Optional[Any]=0.1, a_: int=512, a_: Optional[int]=2, a_: List[str]=0.02, a_: Optional[Any]=1E-12, a_: Union[str, Any]=256, a_: Optional[int]=10, a_: Tuple=1E-3, a_: str=5, a_: Optional[int]=320, a_: str=13_353_718, a_: Dict=5_000, a_: Any=1, a_: List[str]=0, a_: Tuple=2, **a_: Dict, ): '''simple docstring''' super().__init__(pad_token_id=a_, bos_token_id=a_, eos_token_id=a_, **a_ ) # Common config _snake_case : Tuple = vocab_size _snake_case : Optional[int] = max_position_embeddings _snake_case : Any = hidden_size _snake_case : str = retriever_proj_size _snake_case : Optional[int] = num_hidden_layers _snake_case : Dict = num_attention_heads _snake_case : List[Any] = num_candidates _snake_case : List[str] = intermediate_size _snake_case : Any = hidden_act _snake_case : int = hidden_dropout_prob _snake_case : Optional[int] = attention_probs_dropout_prob _snake_case : Optional[Any] = initializer_range _snake_case : Optional[Any] = type_vocab_size _snake_case : Optional[Any] = layer_norm_eps # Reader config _snake_case : Tuple = span_hidden_size _snake_case : List[Any] = max_span_width _snake_case : Dict = reader_layer_norm_eps _snake_case : Dict = reader_beam_size _snake_case : int = reader_seq_len # Retrieval config _snake_case : Optional[int] = num_block_records _snake_case : int = searcher_beam_size
350
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : Union[str, Any] = [] _snake_case : Dict = set({"""(""", """[""", """{"""} ) _snake_case : Union[str, Any] = set({""")""", """]""", """}"""} ) _snake_case : Tuple = {"""{""": """}""", """[""": """]""", """(""": """)"""} for i in range(len(snake_case__ ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(snake_case__ ) == 0 or (len(snake_case__ ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(snake_case__ ) == 0 def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = input("""Enter sequence of brackets: """ ) if is_balanced(snake_case__ ): print(snake_case__ , """is balanced""" ) else: print(snake_case__ , """is not balanced""" ) if __name__ == "__main__": main()
132
0
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import evaluate import numpy as np from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/text-classification/requirements.txt""") __a: Union[str, Any] = logging.getLogger(__name__) @dataclass class UpperCAmelCase : '''simple docstring''' SCREAMING_SNAKE_CASE = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={ "help": ( "Whether to pad all samples to `max_seq_length`. " "If False, will pad the samples dynamically when batching to the maximum length in the batch." ) } , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of prediction examples to this " "value if set." ) } , ) @dataclass class UpperCAmelCase : '''simple docstring''' SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Evaluation language. Also train language if `train_language` is set to None."} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Train language if it is different from the evaluation language."} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "arg to indicate if tokenizer should do lower case in AutoTokenizer.from_pretrained()"} , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) SCREAMING_SNAKE_CASE = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) SCREAMING_SNAKE_CASE = field( default=a__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase__ : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : Tuple = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_xnli''' , UpperCAmelCase ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowercase__ : List[Any] = training_args.get_process_log_level() logger.setLevel(UpperCAmelCase ) datasets.utils.logging.set_verbosity(UpperCAmelCase ) transformers.utils.logging.set_verbosity(UpperCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowercase__ : Any = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase__ : List[str] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # In distributed training, the load_dataset function guarantees that only one local process can concurrently # download the dataset. # Downloading and loading xnli dataset from the hub. if training_args.do_train: if model_args.train_language is None: lowercase__ : Optional[Any] = load_dataset( '''xnli''' , model_args.language , split='''train''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: lowercase__ : Union[str, Any] = load_dataset( '''xnli''' , model_args.train_language , split='''train''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) lowercase__ : Any = train_dataset.features['''label'''].names if training_args.do_eval: lowercase__ : str = load_dataset( '''xnli''' , model_args.language , split='''validation''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) lowercase__ : Optional[int] = eval_dataset.features['''label'''].names if training_args.do_predict: lowercase__ : Optional[Any] = load_dataset( '''xnli''' , model_args.language , split='''test''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) lowercase__ : Optional[int] = predict_dataset.features['''label'''].names # Labels lowercase__ : str = len(UpperCAmelCase ) # Load pretrained model and tokenizer # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase__ : Any = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=UpperCAmelCase , idalabel={str(UpperCAmelCase ): label for i, label in enumerate(UpperCAmelCase )} , labelaid={label: i for i, label in enumerate(UpperCAmelCase )} , finetuning_task='''xnli''' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowercase__ : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , do_lower_case=model_args.do_lower_case , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowercase__ : List[Any] = AutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=UpperCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # Preprocessing the datasets # Padding strategy if data_args.pad_to_max_length: lowercase__ : int = '''max_length''' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch lowercase__ : Any = False def preprocess_function(UpperCAmelCase ): # Tokenize the texts return tokenizer( examples['''premise'''] , examples['''hypothesis'''] , padding=UpperCAmelCase , max_length=data_args.max_seq_length , truncation=UpperCAmelCase , ) if training_args.do_train: if data_args.max_train_samples is not None: lowercase__ : str = min(len(UpperCAmelCase ) , data_args.max_train_samples ) lowercase__ : Optional[Any] = train_dataset.select(range(UpperCAmelCase ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): lowercase__ : Optional[int] = train_dataset.map( UpperCAmelCase , batched=UpperCAmelCase , load_from_cache_file=not data_args.overwrite_cache , desc='''Running tokenizer on train dataset''' , ) # Log a few random samples from the training set: for index in random.sample(range(len(UpperCAmelCase ) ) , 3 ): logger.info(F"""Sample {index} of the training set: {train_dataset[index]}.""" ) if training_args.do_eval: if data_args.max_eval_samples is not None: lowercase__ : Union[str, Any] = min(len(UpperCAmelCase ) , data_args.max_eval_samples ) lowercase__ : Union[str, Any] = eval_dataset.select(range(UpperCAmelCase ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): lowercase__ : Optional[int] = eval_dataset.map( UpperCAmelCase , batched=UpperCAmelCase , load_from_cache_file=not data_args.overwrite_cache , desc='''Running tokenizer on validation dataset''' , ) if training_args.do_predict: if data_args.max_predict_samples is not None: lowercase__ : str = min(len(UpperCAmelCase ) , data_args.max_predict_samples ) lowercase__ : Union[str, Any] = predict_dataset.select(range(UpperCAmelCase ) ) with training_args.main_process_first(desc='''prediction dataset map pre-processing''' ): lowercase__ : Tuple = predict_dataset.map( UpperCAmelCase , batched=UpperCAmelCase , load_from_cache_file=not data_args.overwrite_cache , desc='''Running tokenizer on prediction dataset''' , ) # Get the metric function lowercase__ : Optional[int] = evaluate.load('''xnli''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(UpperCAmelCase ): lowercase__ : Dict = p.predictions[0] if isinstance(p.predictions , UpperCAmelCase ) else p.predictions lowercase__ : int = np.argmax(UpperCAmelCase , axis=1 ) return metric.compute(predictions=UpperCAmelCase , references=p.label_ids ) # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: lowercase__ : Any = default_data_collator elif training_args.fpaa: lowercase__ : int = DataCollatorWithPadding(UpperCAmelCase , pad_to_multiple_of=8 ) else: lowercase__ : List[Any] = None # Initialize our Trainer lowercase__ : Optional[Any] = Trainer( model=UpperCAmelCase , args=UpperCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=UpperCAmelCase , tokenizer=UpperCAmelCase , data_collator=UpperCAmelCase , ) # Training if training_args.do_train: lowercase__ : Union[str, Any] = None if training_args.resume_from_checkpoint is not None: lowercase__ : Any = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowercase__ : List[str] = last_checkpoint lowercase__ : Dict = trainer.train(resume_from_checkpoint=UpperCAmelCase ) lowercase__ : int = train_result.metrics lowercase__ : Tuple = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCAmelCase ) ) lowercase__ : Tuple = min(UpperCAmelCase , len(UpperCAmelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('''train''' , UpperCAmelCase ) trainer.save_metrics('''train''' , UpperCAmelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase__ : str = trainer.evaluate(eval_dataset=UpperCAmelCase ) lowercase__ : Optional[int] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(UpperCAmelCase ) lowercase__ : Optional[int] = min(UpperCAmelCase , len(UpperCAmelCase ) ) trainer.log_metrics('''eval''' , UpperCAmelCase ) trainer.save_metrics('''eval''' , UpperCAmelCase ) # Prediction if training_args.do_predict: logger.info('''*** Predict ***''' ) lowercase__ , lowercase__ , lowercase__ : int = trainer.predict(UpperCAmelCase , metric_key_prefix='''predict''' ) lowercase__ : str = ( data_args.max_predict_samples if data_args.max_predict_samples is not None else len(UpperCAmelCase ) ) lowercase__ : List[Any] = min(UpperCAmelCase , len(UpperCAmelCase ) ) trainer.log_metrics('''predict''' , UpperCAmelCase ) trainer.save_metrics('''predict''' , UpperCAmelCase ) lowercase__ : Tuple = np.argmax(UpperCAmelCase , axis=1 ) lowercase__ : Optional[int] = os.path.join(training_args.output_dir , '''predictions.txt''' ) if trainer.is_world_process_zero(): with open(UpperCAmelCase , '''w''' ) as writer: writer.write('''index\tprediction\n''' ) for index, item in enumerate(UpperCAmelCase ): lowercase__ : Optional[int] = label_list[item] writer.write(F"""{index}\t{item}\n""" ) if __name__ == "__main__": main()
198
'''simple docstring''' from __future__ import annotations from random import random class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase = None ) -> str: lowercase__ : Any = value lowercase__ : Union[str, Any] = random() lowercase__ : Node | None = None lowercase__ : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return F"""'{self.value}: {self.prior:.5}'""" else: return pformat( {F"""{self.value}: {self.prior:.5}""": (self.left, self.right)} , indent=1 ) def __str__( self ) -> str: lowercase__ : Any = str(self.value ) + ''' ''' lowercase__ : str = str(self.left or '''''' ) lowercase__ : List[str] = str(self.right or '''''' ) return value + left + right def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase__ , lowercase__ : int = split(root.left , UpperCAmelCase ) return left, root else: lowercase__ , lowercase__ : Optional[Any] = split(root.right , UpperCAmelCase ) return root, right def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase__ : Any = merge(left.right , UpperCAmelCase ) return left else: lowercase__ : Dict = merge(UpperCAmelCase , right.left ) return right def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): lowercase__ : Dict = Node(UpperCAmelCase ) lowercase__ , lowercase__ : Tuple = split(UpperCAmelCase , UpperCAmelCase ) return merge(merge(UpperCAmelCase , UpperCAmelCase ) , UpperCAmelCase ) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): lowercase__ , lowercase__ : Dict = split(UpperCAmelCase , value - 1 ) lowercase__ , lowercase__ : int = split(UpperCAmelCase , UpperCAmelCase ) return merge(UpperCAmelCase , UpperCAmelCase ) def __UpperCamelCase ( UpperCAmelCase ): if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): for arg in args.split(): if arg[0] == "+": lowercase__ : Tuple = insert(UpperCAmelCase , int(arg[1:] ) ) elif arg[0] == "-": lowercase__ : Union[str, Any] = erase(UpperCAmelCase , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def __UpperCamelCase ( ): lowercase__ : Tuple = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase__ : Dict = input() while args != "q": lowercase__ : Any = interact_treap(UpperCAmelCase , UpperCAmelCase ) print(UpperCAmelCase ) lowercase__ : List[Any] = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
198
1
from torch import nn class _lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Any: '''simple docstring''' super().__init__() A_ : Optional[Any] = class_size A_ : List[str] = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) A_ : Dict = nn.Linear(lowercase_ , lowercase_ ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' A_ : Any = self.mlp(lowercase_ ) return logits
371
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase = [ """python""", """tqdm""", """regex""", """requests""", """packaging""", """filelock""", """numpy""", """tokenizers""", """huggingface-hub""", """safetensors""", """accelerate""", """pyyaml""", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): require_version(deps[pkg] , SCREAMING_SNAKE_CASE )
65
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase = {'''tokenization_bertweet''': ['''BertweetTokenizer''']} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) A =logging.getLogger(__name__) A ='Hello world! cécé herlolip' A =namedtuple( 'BertAbsConfig', [ 'temp_dir', 'large', 'use_bert_emb', 'finetune_bert', 'encoder', 'share_emb', 'max_pos', 'enc_layers', 'enc_hidden_size', 'enc_heads', 'enc_ff_size', 'enc_dropout', 'dec_layers', 'dec_hidden_size', 'dec_heads', 'dec_ff_size', 'dec_dropout', ], ) def snake_case_ (_a : List[Any] , _a : Any ): UpperCAmelCase = BertAbsConfig( temp_dir='''.''' , finetune_bert=_a , large=_a , share_emb=_a , use_bert_emb=_a , encoder='''bert''' , max_pos=5_1_2 , enc_layers=6 , enc_hidden_size=5_1_2 , enc_heads=8 , enc_ff_size=5_1_2 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=7_6_8 , dec_heads=8 , dec_ff_size=2_0_4_8 , dec_dropout=0.2 , ) UpperCAmelCase = torch.load(_a , lambda _a , _a : storage ) UpperCAmelCase = AbsSummarizer(_a , torch.device('''cpu''' ) , _a ) original.eval() UpperCAmelCase = BertAbsSummarizer(_a , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) UpperCAmelCase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs UpperCAmelCase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(_a )) ) UpperCAmelCase = torch.tensor(_a ).unsqueeze(0 ) UpperCAmelCase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(_a )) ) UpperCAmelCase = torch.tensor(_a ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass UpperCAmelCase = encoder_input_ids UpperCAmelCase = decoder_input_ids UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical UpperCAmelCase = original(_a , _a , _a , _a , _a , _a , _a )[0] UpperCAmelCase = original.generator(_a ) UpperCAmelCase = new_model( _a , _a , _a , _a , _a )[0] UpperCAmelCase = new_model.generator(_a ) UpperCAmelCase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(_a ) ) UpperCAmelCase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(_a ) ) UpperCAmelCase = torch.allclose(_a , _a , atol=1E-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": A =argparse.ArgumentParser() parser.add_argument( '--bertabs_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.', ) A =parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
34
0
'''simple docstring''' import numpy as np def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 1e-12 , lowerCAmelCase = 1_00 , ): """simple docstring""" assert np.shape(lowerCAmelCase )[0] == np.shape(lowerCAmelCase )[1] # Ensure proper dimensionality. assert np.shape(lowerCAmelCase )[0] == np.shape(lowerCAmelCase )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowerCAmelCase ) == np.iscomplexobj(lowerCAmelCase ) _lowerCAmelCase = np.iscomplexobj(lowerCAmelCase ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowerCAmelCase , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCAmelCase = False _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCAmelCase = np.dot(lowerCAmelCase , lowerCAmelCase ) # Normalize the resulting output vector. _lowerCAmelCase = w / np.linalg.norm(lowerCAmelCase ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCAmelCase = vector.conj().T if is_complex else vector.T _lowerCAmelCase = np.dot(lowerCAmelCase , np.dot(lowerCAmelCase , lowerCAmelCase ) ) # Check convergence. _lowerCAmelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCAmelCase = True _lowerCAmelCase = lambda_ if is_complex: _lowerCAmelCase = np.real(lambda_ ) return lambda_, vector def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCAmelCase = np.array([41, 4, 20] ) _lowerCAmelCase = real_input_matrix.astype(np.complexaaa ) _lowerCAmelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCAmelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCAmelCase = real_input_matrix _lowerCAmelCase = real_vector elif problem_type == "complex": _lowerCAmelCase = complex_input_matrix _lowerCAmelCase = complex_vector # Our implementation. _lowerCAmelCase , _lowerCAmelCase = power_iteration(lowerCAmelCase , lowerCAmelCase ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCAmelCase , _lowerCAmelCase = np.linalg.eigh(lowerCAmelCase ) # Last eigenvalue is the maximum one. _lowerCAmelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCAmelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowerCAmelCase ) - np.abs(lowerCAmelCase ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
220
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase = 4_00_00_00 ): """simple docstring""" _lowerCAmelCase = [] _lowerCAmelCase , _lowerCAmelCase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = b, a + b return sum(lowerCAmelCase ) if __name__ == "__main__": print(F"""{solution() = }""")
220
1
"""simple docstring""" from __future__ import annotations def lowercase (SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : list[list[str]] , SCREAMING_SNAKE_CASE_ : int , ) -> None: SCREAMING_SNAKE_CASE = len(SCREAMING_SNAKE_CASE_ ) # If row is equal to the size of the board it means there are a queen in each row in # the current board (possible_board) if row == n: # We convert the variable possible_board that looks like this: [1, 3, 0, 2] to # this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . '] boards.append(['. ' * i + 'Q ' + '. ' * (n - 1 - i) for i in possible_board] ) return # We iterate each column in the row to find all possible results in each row for col in range(SCREAMING_SNAKE_CASE_ ): # We apply that we learned previously. First we check that in the current board # (possible_board) there are not other same value because if there is it means # that there are a collision in vertical. Then we apply the two formulas we # learned before: # # 45º: y - x = b or 45: row - col = b # 135º: y + x = b or row + col = b. # # And we verify if the results of this two formulas not exist in their variables # respectively. (diagonal_right_collisions, diagonal_left_collisions) # # If any or these are True it means there is a collision so we continue to the # next value in the for loop. if ( col in possible_board or row - col in diagonal_right_collisions or row + col in diagonal_left_collisions ): continue # If it is False we call dfs function again and we update the inputs depth_first_search( [*possible_board, col] , [*diagonal_right_collisions, row - col] , [*diagonal_left_collisions, row + col] , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> None: SCREAMING_SNAKE_CASE = [] depth_first_search([] , [] , [] , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Print all the boards for board in boards: for column in board: print(SCREAMING_SNAKE_CASE_ ) print('' ) print(len(SCREAMING_SNAKE_CASE_ ) , 'solutions were found.' ) if __name__ == "__main__": import doctest doctest.testmod() n_queens_solution(4)
113
"""simple docstring""" import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel __UpperCamelCase = { '''gwf-440k''': { '''url''': '''https://model-server.zqevans2.workers.dev/gwf-440k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 65536, }, '''jmann-small-190k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 65536, }, '''jmann-large-580k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 131072, }, '''maestro-uncond-150k''': { '''url''': '''https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, '''unlocked-uncond-250k''': { '''url''': '''https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, '''honk-140k''': { '''url''': '''https://model-server.zqevans2.workers.dev/honk-140k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, } def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Optional[Any]: return torch.atana(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) / math.pi * 2 def lowercase (SCREAMING_SNAKE_CASE_ : Optional[int] ) -> int: SCREAMING_SNAKE_CASE = torch.sin(t * math.pi / 2 ) ** 2 SCREAMING_SNAKE_CASE = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' pass class lowerCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self , lowerCAmelCase__ ) -> int: super().__init__() SCREAMING_SNAKE_CASE = DiffusionAttnUnetaD(lowerCAmelCase__ , n_attn_layers=4 ) SCREAMING_SNAKE_CASE = deepcopy(self.diffusion ) SCREAMING_SNAKE_CASE = torch.quasirandom.SobolEngine(1 , scramble=lowerCAmelCase__ ) def lowercase (SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> Any: SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['url'] os.system(F'wget {url} ./' ) return F'./{model_name}.ckpt' __UpperCamelCase = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', } __UpperCamelCase = { '''8''': '''resnets.0''', '''9''': '''attentions.0''', '''10''': '''resnets.1''', '''11''': '''attentions.1''', '''12''': '''resnets.2''', '''13''': '''attentions.2''', } __UpperCamelCase = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', '''8''': '''resnets.3''', '''9''': '''attentions.3''', '''10''': '''resnets.4''', '''11''': '''attentions.4''', '''12''': '''resnets.5''', '''13''': '''attentions.5''', } __UpperCamelCase = { '''0''': '''resnets.0''', '''1''': '''resnets.1''', '''2''': '''resnets.2''', '''4''': '''resnets.0''', '''5''': '''resnets.1''', '''6''': '''resnets.2''', } __UpperCamelCase = { '''skip''': '''conv_skip''', '''main.0''': '''conv_1''', '''main.1''': '''group_norm_1''', '''main.3''': '''conv_2''', '''main.4''': '''group_norm_2''', } __UpperCamelCase = { '''norm''': '''group_norm''', '''qkv_proj''': ['''query''', '''key''', '''value'''], '''out_proj''': ['''proj_attn'''], } def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> List[Any]: if name.startswith('skip' ): return name.replace('skip' , RES_CONV_MAP['skip'] ) # name has to be of format main.{digit} if not name.startswith('main.' ): raise ValueError(F'ResConvBlock error with {name}' ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Dict: for key, value in ATTN_MAP.items(): if name.startswith(SCREAMING_SNAKE_CASE_ ) and not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) elif name.startswith(SCREAMING_SNAKE_CASE_ ): return [name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for v in value] raise ValueError(F'Attn error with {name}' ) def lowercase (SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Tuple=13 ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = input_string if string.split('.' )[0] == "timestep_embed": return string.replace('timestep_embed' , 'time_proj' ) SCREAMING_SNAKE_CASE = 0 if string.startswith('net.3.' ): depth += 1 SCREAMING_SNAKE_CASE = string[6:] elif string.startswith('net.' ): SCREAMING_SNAKE_CASE = string[4:] while string.startswith('main.7.' ): depth += 1 SCREAMING_SNAKE_CASE = string[7:] if string.startswith('main.' ): SCREAMING_SNAKE_CASE = string[5:] # mid block if string[:2].isdigit(): SCREAMING_SNAKE_CASE = string[:2] SCREAMING_SNAKE_CASE = string[2:] else: SCREAMING_SNAKE_CASE = string[0] SCREAMING_SNAKE_CASE = string[1:] if depth == max_depth: SCREAMING_SNAKE_CASE = MID_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = 'mid_block' elif depth > 0 and int(SCREAMING_SNAKE_CASE_ ) < 7: SCREAMING_SNAKE_CASE = DOWN_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'down_blocks.{depth}' elif depth > 0 and int(SCREAMING_SNAKE_CASE_ ) > 7: SCREAMING_SNAKE_CASE = UP_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'up_blocks.{max_depth - depth - 1}' elif depth == 0: SCREAMING_SNAKE_CASE = DEPTH_0_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'up_blocks.{max_depth - 1}' if int(SCREAMING_SNAKE_CASE_ ) > 3 else 'down_blocks.0' if not string_left.startswith('.' ): raise ValueError(F'Naming error with {input_string} and string_left: {string_left}.' ) SCREAMING_SNAKE_CASE = string_left[1:] if "resnets" in new_layer: SCREAMING_SNAKE_CASE = convert_resconv_naming(SCREAMING_SNAKE_CASE_ ) elif "attentions" in new_layer: SCREAMING_SNAKE_CASE = convert_attn_naming(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = new_string_left if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = prefix + '.' + new_layer + '.' + string_left else: SCREAMING_SNAKE_CASE = [prefix + '.' + new_layer + '.' + s for s in string_left] return new_string def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Dict: SCREAMING_SNAKE_CASE = {} for k, v in state_dict.items(): if k.endswith('kernel' ): # up- and downsample layers, don't have trainable weights continue SCREAMING_SNAKE_CASE = rename(SCREAMING_SNAKE_CASE_ ) # check if we need to transform from Conv => Linear for attention if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = transform_conv_attns(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else: SCREAMING_SNAKE_CASE = v return new_state_dict def lowercase (SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> str: if len(SCREAMING_SNAKE_CASE_ ) == 1: if len(v.shape ) == 3: # weight SCREAMING_SNAKE_CASE = v[:, :, 0] else: # bias SCREAMING_SNAKE_CASE = v else: # qkv matrices SCREAMING_SNAKE_CASE = v.shape[0] SCREAMING_SNAKE_CASE = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape, :, 0] else: SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def lowercase (SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> List[Any]: SCREAMING_SNAKE_CASE = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) SCREAMING_SNAKE_CASE = args.model_path.split('/' )[-1].split('.' )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), F'Make sure to provide one of the official model names {MODELS_MAP.keys()}' SCREAMING_SNAKE_CASE = download(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['sample_rate'] SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['sample_size'] SCREAMING_SNAKE_CASE = Object() SCREAMING_SNAKE_CASE = sample_size SCREAMING_SNAKE_CASE = sample_rate SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = UNetaDModel(sample_size=SCREAMING_SNAKE_CASE_ , sample_rate=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = diffusers_model.state_dict() SCREAMING_SNAKE_CASE = DiffusionUncond(SCREAMING_SNAKE_CASE_ ) orig_model.load_state_dict(torch.load(args.model_path , map_location=SCREAMING_SNAKE_CASE_ )['state_dict'] ) SCREAMING_SNAKE_CASE = orig_model.diffusion_ema.eval() SCREAMING_SNAKE_CASE = orig_model.state_dict() SCREAMING_SNAKE_CASE = rename_orig_weights(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) SCREAMING_SNAKE_CASE = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(SCREAMING_SNAKE_CASE_ ) == 0, F'Problem with {renamed_minus_diffusers}' assert all(k.endswith('kernel' ) for k in list(SCREAMING_SNAKE_CASE_ ) ), F'Problem with {diffusers_minus_renamed}' for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), F'Shape for {key} doesn\'t match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}' if key == "time_proj.weight": SCREAMING_SNAKE_CASE = value.squeeze() SCREAMING_SNAKE_CASE = value diffusers_model.load_state_dict(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = 1_00 SCREAMING_SNAKE_CASE = 33 SCREAMING_SNAKE_CASE = IPNDMScheduler(num_train_timesteps=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.randn([1, 2, config.sample_size] , generator=SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.linspace(1 , 0 , steps + 1 , device=SCREAMING_SNAKE_CASE_ )[:-1] SCREAMING_SNAKE_CASE = get_crash_schedule(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = DanceDiffusionPipeline(unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE = pipe(num_inference_steps=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ ).audios SCREAMING_SNAKE_CASE = sampling.iplms_sample(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , {} ) SCREAMING_SNAKE_CASE = generated.clamp(-1 , 1 ) SCREAMING_SNAKE_CASE = (generated - audio).abs().sum() SCREAMING_SNAKE_CASE = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print('Diff sum' , SCREAMING_SNAKE_CASE_ ) print('Diff max' , SCREAMING_SNAKE_CASE_ ) assert diff_max < 1E-3, F'Diff max: {diff_max} is too much :-/' print(F'Conversion for {model_name} successful!' ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_path''', default=None, type=str, required=True, help='''Path to the model to convert.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument('''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the output model.''') __UpperCamelCase = parser.parse_args() main(args)
113
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase ( __lowerCAmelCase ): snake_case_ = ['''input_features''', '''attention_mask'''] def __init__( self, lowercase_=80, lowercase_=16000, lowercase_=80, lowercase_=0.0, lowercase_=True, lowercase_=True, lowercase_=True, **lowercase_, ) -> Optional[Any]: super().__init__(feature_size=lowercase_, sampling_rate=lowercase_, padding_value=lowercase_, **lowercase_ ) snake_case = num_mel_bins snake_case = do_ceptral_normalize snake_case = normalize_means snake_case = normalize_vars snake_case = True def _lowerCamelCase ( self, lowercase_, ) -> np.ndarray: snake_case = waveform * (2**15) # Kaldi compliance: 16-bit signed integers snake_case = torch.from_numpy(lowercase_ ).unsqueeze(0 ) snake_case = ta_kaldi.fbank(lowercase_, num_mel_bins=self.num_mel_bins, sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( lowercase_, lowercase_, lowercase_ = True, lowercase_ = True, lowercase_ = 0.0, ) -> np.ndarray: # make sure we normalize float32 arrays if normalize_means: snake_case = x[:input_length].mean(axis=0 ) snake_case = np.subtract(lowercase_, lowercase_ ) if normalize_vars: snake_case = x[:input_length].std(axis=0 ) snake_case = np.divide(lowercase_, lowercase_ ) if input_length < x.shape[0]: snake_case = padding_value # make sure array is in float32 snake_case = x.astype(np.floataa ) return x def _lowerCamelCase ( self, lowercase_, lowercase_ = None ) -> List[np.ndarray]: snake_case = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(lowercase_, lowercase_, self.normalize_means, self.normalize_vars, self.padding_value ) for x, n in zip(lowercase_, lowercase_ ) ] def __call__( self, lowercase_, lowercase_ = False, lowercase_ = None, lowercase_ = False, lowercase_ = None, lowercase_ = None, lowercase_ = None, lowercase_ = None, **lowercase_, ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) snake_case = isinstance(lowercase_, np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) snake_case = is_batched_numpy or ( isinstance(lowercase_, (list, tuple) ) and (isinstance(raw_speech[0], (np.ndarray, tuple, list) )) ) if is_batched: snake_case = [np.asarray(lowercase_, dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowercase_, np.ndarray ): snake_case = np.asarray(lowercase_, dtype=np.floataa ) elif isinstance(lowercase_, np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): snake_case = raw_speech.astype(np.floataa ) # always return batch if not is_batched: snake_case = [raw_speech] # extract fbank features snake_case = [self._extract_fbank_features(lowercase_ ) for waveform in raw_speech] # convert into correct format for padding snake_case = BatchFeature({'input_features': features} ) snake_case = self.pad( lowercase_, padding=lowercase_, max_length=lowercase_, truncation=lowercase_, pad_to_multiple_of=lowercase_, return_attention_mask=lowercase_, **lowercase_, ) # make sure list is in array format snake_case = padded_inputs.get('input_features' ) if isinstance(input_features[0], lowercase_ ): snake_case = [np.asarray(lowercase_, dtype=np.floataa ) for feature in input_features] snake_case = padded_inputs.get('attention_mask' ) if attention_mask is not None: snake_case = [np.asarray(lowercase_, dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: snake_case = ( np.array(lowercase_, dtype=np.intaa ) if self._get_padding_strategies(lowercase_, max_length=lowercase_ ) is not PaddingStrategy.DO_NOT_PAD else None ) snake_case = self.normalize( padded_inputs['input_features'], attention_mask=lowercase_ ) if return_tensors is not None: snake_case = padded_inputs.convert_to_tensors(lowercase_ ) return padded_inputs
332
'''simple docstring''' from __future__ import annotations def __magic_name__ ( A ) -> list: if len(A ) == 0: return [] snake_case , snake_case = min(A ), max(A ) snake_case = int(max_value - min_value ) + 1 snake_case = [[] for _ in range(A )] for i in my_list: buckets[int(i - min_value )].append(A ) return [v for bucket in buckets for v in sorted(A )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -1_0, 1_5, 2, -2]) == [-1_0, -2, 0, 1, 2, 1_5]
332
1
'''simple docstring''' import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) lowercase__ : List[str] = logging.getLogger(__name__) def a__ ( lowercase : Union[str, Any], lowercase : List[str] ) -> Any: """simple docstring""" _UpperCamelCase = np.argmax(a__, axis=1 ) return np.sum(outputs == labels ) def a__ ( lowercase : Union[str, Any] ) -> List[str]: """simple docstring""" with open(a__, encoding='''utf_8''' ) as f: _UpperCamelCase = csv.reader(a__ ) _UpperCamelCase = [] next(a__ ) # skip the first line for line in tqdm(a__ ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( lowercase : str, lowercase : Any, lowercase : Union[str, Any], lowercase : Tuple, lowercase : List[Any], lowercase : Dict ) -> str: """simple docstring""" _UpperCamelCase = [] for dataset in encoded_datasets: _UpperCamelCase = len(a__ ) _UpperCamelCase = np.zeros((n_batch, 2, input_len), dtype=np.intaa ) _UpperCamelCase = np.zeros((n_batch, 2), dtype=np.intaa ) _UpperCamelCase = np.full((n_batch, 2, input_len), fill_value=-100, dtype=np.intaa ) _UpperCamelCase = np.zeros((n_batch,), dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(a__ ): _UpperCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] _UpperCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] _UpperCamelCase = with_conta _UpperCamelCase = with_conta _UpperCamelCase = len(a__ ) - 1 _UpperCamelCase = len(a__ ) - 1 _UpperCamelCase = with_conta _UpperCamelCase = with_conta _UpperCamelCase = mc_label _UpperCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(a__ ) for t in all_inputs ) ) return tensor_datasets def a__ ( ) -> int: """simple docstring""" _UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''', type=a__, default='''openai-gpt''', help='''pretrained model name''' ) parser.add_argument('''--do_train''', action='''store_true''', help='''Whether to run training.''' ) parser.add_argument('''--do_eval''', action='''store_true''', help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''', default=a__, type=a__, required=a__, help='''The output directory where the model predictions and checkpoints will be written.''', ) parser.add_argument('''--train_dataset''', type=a__, default='''''' ) parser.add_argument('''--eval_dataset''', type=a__, default='''''' ) parser.add_argument('''--seed''', type=a__, default=42 ) parser.add_argument('''--num_train_epochs''', type=a__, default=3 ) parser.add_argument('''--train_batch_size''', type=a__, default=8 ) parser.add_argument('''--eval_batch_size''', type=a__, default=16 ) parser.add_argument('''--adam_epsilon''', default=1e-8, type=a__, help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''', type=a__, default=1 ) parser.add_argument( '''--max_steps''', default=-1, type=a__, help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ), ) parser.add_argument( '''--gradient_accumulation_steps''', type=a__, default=1, help='''Number of updates steps to accumulate before performing a backward/update pass.''', ) parser.add_argument('''--learning_rate''', type=a__, default=6.25e-5 ) parser.add_argument('''--warmup_steps''', default=0, type=a__, help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''', type=a__, default='''warmup_linear''' ) parser.add_argument('''--weight_decay''', type=a__, default=0.0_1 ) parser.add_argument('''--lm_coef''', type=a__, default=0.9 ) parser.add_argument('''--n_valid''', type=a__, default=374 ) parser.add_argument('''--server_ip''', type=a__, default='''''', help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''', type=a__, default='''''', help='''Can be used for distant debugging.''' ) _UpperCamelCase = parser.parse_args() print(a__ ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port), redirect_output=a__ ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) _UpperCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) _UpperCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(a__, a__ ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset _UpperCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] _UpperCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(a__ ) _UpperCamelCase = tokenizer.convert_tokens_to_ids(a__ ) _UpperCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(a__ ) ) model.to(a__ ) # Load and encode the datasets def tokenize_and_encode(lowercase : Optional[int] ): if isinstance(a__, a__ ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(a__ ) ) elif isinstance(a__, a__ ): return obj return [tokenize_and_encode(a__ ) for o in obj] logger.info('''Encoding dataset...''' ) _UpperCamelCase = load_rocstories_dataset(args.train_dataset ) _UpperCamelCase = load_rocstories_dataset(args.eval_dataset ) _UpperCamelCase = (train_dataset, eval_dataset) _UpperCamelCase = tokenize_and_encode(a__ ) # Compute the max input length for the Transformer _UpperCamelCase = model.config.n_positions // 2 - 2 _UpperCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ), len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) _UpperCamelCase = min(a__, model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders _UpperCamelCase = pre_process_datasets(a__, a__, a__, *a__ ) _UpperCamelCase , _UpperCamelCase = tensor_datasets[0], tensor_datasets[1] _UpperCamelCase = TensorDataset(*a__ ) _UpperCamelCase = RandomSampler(a__ ) _UpperCamelCase = DataLoader(a__, sampler=a__, batch_size=args.train_batch_size ) _UpperCamelCase = TensorDataset(*a__ ) _UpperCamelCase = SequentialSampler(a__ ) _UpperCamelCase = DataLoader(a__, sampler=a__, batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: _UpperCamelCase = args.max_steps _UpperCamelCase = args.max_steps // (len(a__ ) // args.gradient_accumulation_steps) + 1 else: _UpperCamelCase = len(a__ ) // args.gradient_accumulation_steps * args.num_train_epochs _UpperCamelCase = list(model.named_parameters() ) _UpperCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] _UpperCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] _UpperCamelCase = AdamW(a__, lr=args.learning_rate, eps=args.adam_epsilon ) _UpperCamelCase = get_linear_schedule_with_warmup( a__, num_warmup_steps=args.warmup_steps, num_training_steps=a__ ) if args.do_train: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ), desc='''Epoch''' ): _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(a__, desc='''Training''' ) for step, batch in enumerate(a__ ): _UpperCamelCase = tuple(t.to(a__ ) for t in batch ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = batch _UpperCamelCase = model(a__, mc_token_ids=a__, lm_labels=a__, mc_labels=a__ ) _UpperCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() _UpperCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 _UpperCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(a__, scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer _UpperCamelCase = model.module if hasattr(a__, '''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` _UpperCamelCase = os.path.join(args.output_dir, a__ ) _UpperCamelCase = os.path.join(args.output_dir, a__ ) torch.save(model_to_save.state_dict(), a__ ) model_to_save.config.to_json_file(a__ ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned _UpperCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) _UpperCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(a__ ) if args.do_eval: model.eval() _UpperCamelCase , _UpperCamelCase = 0, 0 _UpperCamelCase , _UpperCamelCase = 0, 0 for batch in tqdm(a__, desc='''Evaluating''' ): _UpperCamelCase = tuple(t.to(a__ ) for t in batch ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = batch with torch.no_grad(): _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = model( a__, mc_token_ids=a__, lm_labels=a__, mc_labels=a__ ) _UpperCamelCase = mc_logits.detach().cpu().numpy() _UpperCamelCase = mc_labels.to('''cpu''' ).numpy() _UpperCamelCase = accuracy(a__, a__ ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 _UpperCamelCase = eval_loss / nb_eval_steps _UpperCamelCase = eval_accuracy / nb_eval_examples _UpperCamelCase = tr_loss / nb_tr_steps if args.do_train else None _UpperCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} _UpperCamelCase = os.path.join(args.output_dir, '''eval_results.txt''' ) with open(a__, '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''', a__, str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
324
"""simple docstring""" def lowercase ( a__ : str ) -> list[int]: _UpperCamelCase = [0 for i in range(len(a__ ) )] # initialize interval's left pointer and right pointer _UpperCamelCase , _UpperCamelCase = 0, 0 for i in range(1 , len(a__ ) ): # case when current index is inside the interval if i <= right_pointer: _UpperCamelCase = min(right_pointer - i + 1 , z_result[i - left_pointer] ) _UpperCamelCase = min_edge while go_next(a__ , a__ , a__ ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: _UpperCamelCase , _UpperCamelCase = i, i + z_result[i] - 1 return z_result def lowercase ( a__ : int , a__ : list[int] , a__ : str ) -> bool: return i + z_result[i] < len(a__ ) and s[z_result[i]] == s[i + z_result[i]] def lowercase ( a__ : str , a__ : str ) -> int: _UpperCamelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string _UpperCamelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(a__ ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
256
0
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup __lowerCamelCase : Optional[int] = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): def __init__( self : List[Any] , **_lowercase : Optional[int] ): """simple docstring""" requires_backends(self , ["""bs4"""] ) super().__init__(**_lowercase ) def __a ( self : Dict , _lowercase : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag SCREAMING_SNAKE_CASE__ = parent.find_all(child.name , recursive=_lowercase ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(_lowercase ) else next(i for i, s in enumerate(_lowercase , 1 ) if s is child ) ) SCREAMING_SNAKE_CASE__ = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def __a ( self : Tuple , _lowercase : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = BeautifulSoup(_lowercase , """html.parser""" ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] for element in html_code.descendants: if type(_lowercase ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue SCREAMING_SNAKE_CASE__ = html.unescape(_lowercase ).strip() if not text_in_this_tag: continue all_doc_strings.append(_lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.xpath_soup(_lowercase ) stringaxtag_seq.append(_lowercase ) stringaxsubs_seq.append(_lowercase ) if len(_lowercase ) != len(_lowercase ): raise ValueError("""Number of doc strings and xtags does not correspond""" ) if len(_lowercase ) != len(_lowercase ): raise ValueError("""Number of doc strings and xsubs does not correspond""" ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def __a ( self : Optional[int] , _lowercase : Dict , _lowercase : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """""" for tagname, subs in zip(_lowercase , _lowercase ): xpath += f"""/{tagname}""" if subs != 0: xpath += f"""[{subs}]""" return xpath def __call__( self : Dict , _lowercase : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = False # Check that strings has a valid type if isinstance(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE__ = True elif isinstance(_lowercase , (list, tuple) ): if len(_lowercase ) == 0 or isinstance(html_strings[0] , _lowercase ): SCREAMING_SNAKE_CASE__ = True if not valid_strings: raise ValueError( """HTML strings must of type `str`, `List[str]` (batch of examples), """ f"""but is of type {type(_lowercase )}.""" ) SCREAMING_SNAKE_CASE__ = bool(isinstance(_lowercase , (list, tuple) ) and (isinstance(html_strings[0] , _lowercase )) ) if not is_batched: SCREAMING_SNAKE_CASE__ = [html_strings] # Get nodes + xpaths SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] for html_string in html_strings: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.get_three_from_single(_lowercase ) nodes.append(_lowercase ) SCREAMING_SNAKE_CASE__ = [] for node, tag_list, sub_list in zip(_lowercase , _lowercase , _lowercase ): SCREAMING_SNAKE_CASE__ = self.construct_xpath(_lowercase , _lowercase ) xpath_strings.append(_lowercase ) xpaths.append(_lowercase ) # return as Dict SCREAMING_SNAKE_CASE__ = {"""nodes""": nodes, """xpaths""": xpaths} SCREAMING_SNAKE_CASE__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) return encoded_inputs
204
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __lowerCamelCase : Any = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = ["input_features", "attention_mask"] def __init__( self : int , _lowercase : Dict=80 , _lowercase : Union[str, Any]=1_60_00 , _lowercase : str=0.0 , _lowercase : int=10 , _lowercase : str=25 , _lowercase : Union[str, Any]="hamming_window" , _lowercase : Dict=3_27_68.0 , _lowercase : Optional[int]=0.97 , _lowercase : Optional[int]=1.0 , _lowercase : int=True , _lowercase : Union[str, Any]=True , _lowercase : Optional[int]=False , **_lowercase : Union[str, Any] , ): """simple docstring""" super().__init__(feature_size=_lowercase , sampling_rate=_lowercase , padding_value=_lowercase , **_lowercase ) SCREAMING_SNAKE_CASE__ = feature_size SCREAMING_SNAKE_CASE__ = sampling_rate SCREAMING_SNAKE_CASE__ = padding_value SCREAMING_SNAKE_CASE__ = hop_length SCREAMING_SNAKE_CASE__ = win_length SCREAMING_SNAKE_CASE__ = frame_signal_scale SCREAMING_SNAKE_CASE__ = preemphasis_coeff SCREAMING_SNAKE_CASE__ = mel_floor SCREAMING_SNAKE_CASE__ = normalize_means SCREAMING_SNAKE_CASE__ = normalize_vars SCREAMING_SNAKE_CASE__ = win_function SCREAMING_SNAKE_CASE__ = return_attention_mask SCREAMING_SNAKE_CASE__ = win_length * sampling_rate // 10_00 SCREAMING_SNAKE_CASE__ = hop_length * sampling_rate // 10_00 SCREAMING_SNAKE_CASE__ = optimal_fft_length(self.sample_size ) SCREAMING_SNAKE_CASE__ = (self.n_fft // 2) + 1 def __a ( self : str , _lowercase : np.array ): """simple docstring""" if self.win_function == "hamming_window": SCREAMING_SNAKE_CASE__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=_lowercase ) else: SCREAMING_SNAKE_CASE__ = window_function(window_length=self.sample_size , name=self.win_function ) SCREAMING_SNAKE_CASE__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) SCREAMING_SNAKE_CASE__ = spectrogram( one_waveform * self.frame_signal_scale , window=_lowercase , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=_lowercase , preemphasis=self.preemphasis_coeff , mel_filters=_lowercase , mel_floor=self.mel_floor , log_mel="""log""" , ) return msfc_features.T def __a ( self : Tuple , _lowercase : Union[str, Any] , _lowercase : Optional[int] , _lowercase : int ): """simple docstring""" if self.normalize_means: SCREAMING_SNAKE_CASE__ = x[:input_length].mean(axis=0 ) SCREAMING_SNAKE_CASE__ = np.subtract(_lowercase , _lowercase ) if self.normalize_vars: SCREAMING_SNAKE_CASE__ = x[:input_length].std(axis=0 ) SCREAMING_SNAKE_CASE__ = np.divide(_lowercase , _lowercase ) if input_length < x.shape[0]: SCREAMING_SNAKE_CASE__ = padding_value # make sure array is in float32 SCREAMING_SNAKE_CASE__ = x.astype(np.floataa ) return x def __a ( self : int , _lowercase : List[np.ndarray] , _lowercase : Optional[np.ndarray] = None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(_lowercase , _lowercase , self.padding_value ) for x, n in zip(_lowercase , _lowercase )] def __call__( self : Dict , _lowercase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , _lowercase : Union[bool, str, PaddingStrategy] = False , _lowercase : Optional[int] = None , _lowercase : bool = False , _lowercase : Optional[int] = None , _lowercase : Optional[bool] = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : Optional[int] = None , **_lowercase : List[str] , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) SCREAMING_SNAKE_CASE__ = isinstance(_lowercase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) SCREAMING_SNAKE_CASE__ = is_batched_numpy or ( isinstance(_lowercase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: SCREAMING_SNAKE_CASE__ = [np.asarray(_lowercase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_lowercase , np.ndarray ): SCREAMING_SNAKE_CASE__ = np.asarray(_lowercase , dtype=np.floataa ) elif isinstance(_lowercase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): SCREAMING_SNAKE_CASE__ = raw_speech.astype(np.floataa ) # always return batch if not is_batched: SCREAMING_SNAKE_CASE__ = [raw_speech] # extract fbank features SCREAMING_SNAKE_CASE__ = [self._extract_mfsc_features(_lowercase ) for one_waveform in raw_speech] # convert into correct format for padding SCREAMING_SNAKE_CASE__ = BatchFeature({"""input_features""": features} ) SCREAMING_SNAKE_CASE__ = self.pad( _lowercase , padding=_lowercase , max_length=_lowercase , truncation=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) # make sure list is in array format SCREAMING_SNAKE_CASE__ = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] , _lowercase ): SCREAMING_SNAKE_CASE__ = [np.asarray(_lowercase , dtype=np.floataa ) for feature in input_features] SCREAMING_SNAKE_CASE__ = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: SCREAMING_SNAKE_CASE__ = [np.asarray(_lowercase , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: SCREAMING_SNAKE_CASE__ = ( np.array(_lowercase , dtype=np.intaa ) if self._get_padding_strategies(_lowercase , max_length=_lowercase ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) SCREAMING_SNAKE_CASE__ = self.normalize( padded_inputs["""input_features"""] , attention_mask=_lowercase ) if return_tensors is not None: SCREAMING_SNAKE_CASE__ = padded_inputs.convert_to_tensors(_lowercase ) return padded_inputs
204
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCamelCase__ ).to(lowerCamelCase__ ) __lowerCamelCase = AutoTokenizer.from_pretrained('google/mt5-small' ) __lowerCamelCase = tokenizer('Hello there' , return_tensors='pt' ).input_ids __lowerCamelCase = tokenizer('Hi I am' , return_tensors='pt' ).input_ids __lowerCamelCase = model(input_ids.to(lowerCamelCase__ ) , labels=labels.to(lowerCamelCase__ ) ).loss __lowerCamelCase = -(labels.shape[-1] * loss.item()) __lowerCamelCase = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
90
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = { 'task_specific_params': { 'summarization': {'length_penalty': 1.0, 'max_length': 128, 'min_length': 12, 'num_beams': 4}, 'summarization_cnn': {'length_penalty': 2.0, 'max_length': 142, 'min_length': 56, 'num_beams': 4}, 'summarization_xsum': {'length_penalty': 1.0, 'max_length': 62, 'min_length': 11, 'num_beams': 6}, } } __lowerCamelCase = { 'task_specific_params.summarization.length_penalty': 1.0, 'task_specific_params.summarization.max_length': 128, 'task_specific_params.summarization.min_length': 12, 'task_specific_params.summarization.num_beams': 4, 'task_specific_params.summarization_cnn.length_penalty': 2.0, 'task_specific_params.summarization_cnn.max_length': 142, 'task_specific_params.summarization_cnn.min_length': 56, 'task_specific_params.summarization_cnn.num_beams': 4, 'task_specific_params.summarization_xsum.length_penalty': 1.0, 'task_specific_params.summarization_xsum.max_length': 62, 'task_specific_params.summarization_xsum.min_length': 11, 'task_specific_params.summarization_xsum.num_beams': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , np.reshape(lowerCamelCase__ , (12, 5) ) ) ) @require_torch def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , reshape(lowerCamelCase__ , (12, 5) ).numpy() ) ) @require_tf def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , reshape(lowerCamelCase__ , (12, 5) ).numpy() ) ) @require_flax def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) __lowerCamelCase = np.random.randn(3 , 4 , 5 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , np.asarray(reshape(lowerCamelCase__ , (12, 5) ) ) ) ) def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) __lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = np.random.randn(1 , 3 , 4 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) __lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = np.random.randn(1 , 3 , 4 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) __lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def lowercase_ ( self ) -> List[str]: '''simple docstring''' __lowerCamelCase = np.random.randn(1 , 3 , 4 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) __lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = np.random.randn(3 , 4 ) __lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
90
1
"""simple docstring""" import operator as op a : List[Any] = '''scaler.pt''' a : Any = '''pytorch_model''' a : Union[str, Any] = '''random_states''' a : List[Any] = '''optimizer''' a : List[str] = '''scheduler''' a : Dict = '''pytorch_model.bin''' a : List[str] = '''pytorch_model.bin.index.json''' a : Optional[int] = '''model.safetensors''' a : List[str] = '''model.safetensors.index.json''' a : Optional[int] = '''1.10.2''' a : Dict = '''py38''' a : Dict = '''4.17.0''' a : Union[str, Any] = ['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] a : int = ['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] a : Tuple = ['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] a : Tuple = ['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] a : List[Any] = ['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] a : Dict = '''2.0.1''' a : Tuple = ['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] a : str = ['''default''', '''reduce-overhead''', '''max-autotune'''] a : Any = {'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 a : List[Any] = [ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] a : str = ['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] a : Dict = ['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
352
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->Union[str, Any]: '''simple docstring''' if isinstance(_lowercase , collections.abc.Iterable ): return x return (x, x) @require_flax class __UpperCamelCase : def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: pass def __a ( self ) -> List[Any]: pass def __a ( self ) -> str: pass def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: a : Dict = np.abs((a - b) ).max() self.assertLessEqual(lowerCAmelCase__ , lowerCAmelCase__ , f"""Difference between torch and flax is {diff} (>= {tol}).""" ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Dict: a : Dict = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = FlaxVisionTextDualEncoderModel(lowerCAmelCase__ ) a : int = model(input_ids=lowerCAmelCase__ , pixel_values=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Optional[Any]: a, a : Optional[int] = self.get_vision_text_model(lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = {"vision_model": vision_model, "text_model": text_model} a : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCAmelCase__ ) a : List[str] = model(input_ids=lowerCAmelCase__ , pixel_values=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Union[str, Any]: a, a : Dict = self.get_vision_text_model(lowerCAmelCase__ , lowerCAmelCase__ ) a : Tuple = {"vision_model": vision_model, "text_model": text_model} a : Tuple = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCAmelCase__ ) a : List[str] = model(input_ids=lowerCAmelCase__ , pixel_values=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) a : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCAmelCase__ ) a : str = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCAmelCase__ ) a : Dict = model(input_ids=lowerCAmelCase__ , pixel_values=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) a : List[Any] = after_output[0] a : Optional[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCAmelCase__ , 1E-3 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> List[Any]: a, a : Union[str, Any] = self.get_vision_text_model(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = {"vision_model": vision_model, "text_model": text_model} a : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCAmelCase__ ) a : Tuple = model( input_ids=lowerCAmelCase__ , pixel_values=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , output_attentions=lowerCAmelCase__ ) a : int = output.vision_model_output.attentions self.assertEqual(len(lowerCAmelCase__ ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) a : Optional[int] = to_atuple(vision_model.config.image_size ) a : Tuple = to_atuple(vision_model.config.patch_size ) a : Any = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) a : Dict = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) a : str = output.text_model_output.attentions self.assertEqual(len(lowerCAmelCase__ ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: pt_model.to(lowerCAmelCase__ ) pt_model.eval() # prepare inputs a : List[Any] = inputs_dict a : Any = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): a : int = pt_model(**lowerCAmelCase__ ).to_tuple() a : Union[str, Any] = fx_model(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(lowerCAmelCase__ , pt_output.numpy() , 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowerCAmelCase__ ) a : Dict = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCAmelCase__ , from_pt=lowerCAmelCase__ ) a : Optional[int] = fx_model_loaded(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(lowerCAmelCase__ , pt_output.numpy() , 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowerCAmelCase__ ) a : Optional[int] = VisionTextDualEncoderModel.from_pretrained(lowerCAmelCase__ , from_flax=lowerCAmelCase__ ) pt_model_loaded.to(lowerCAmelCase__ ) pt_model_loaded.eval() with torch.no_grad(): a : int = pt_model_loaded(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(lowerCAmelCase__ , pt_output_loaded.numpy() , 4E-2 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: a : List[Any] = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = VisionTextDualEncoderModel(lowerCAmelCase__ ) a : Optional[int] = FlaxVisionTextDualEncoderModel(lowerCAmelCase__ ) a : Dict = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , lowerCAmelCase__ ) a : List[str] = fx_state self.check_pt_flax_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCAmelCase__ , lowerCAmelCase__ ) a : Optional[int] = VisionTextDualEncoderModel(lowerCAmelCase__ ) a : List[Any] = FlaxVisionTextDualEncoderModel(lowerCAmelCase__ ) a : int = load_flax_weights_in_pytorch_model(lowerCAmelCase__ , fx_model.params ) self.check_pt_flax_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Dict: a : Any = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowerCAmelCase__ ) def __a ( self ) -> Dict: a : List[str] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : int = self.prepare_config_and_inputs() self.check_save_load(**lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : Tuple = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowerCAmelCase__ ) @is_pt_flax_cross_test def __a ( self ) -> Any: a : List[Any] = self.prepare_config_and_inputs() a : Tuple = config_inputs_dict.pop("vision_config" ) a : int = config_inputs_dict.pop("text_config" ) a : List[str] = config_inputs_dict self.check_equivalence_pt_to_flax(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) self.check_equivalence_flax_to_pt(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) @slow def __a ( self ) -> List[Any]: a, a : Optional[int] = self.get_pretrained_model_and_inputs() a : Optional[int] = model_a(**lowerCAmelCase__ ) a : Optional[int] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowerCAmelCase__ ) a : Any = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCAmelCase__ ) a : str = model_a(**lowerCAmelCase__ ) a : Dict = after_outputs[0] a : Optional[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCAmelCase__ , 1E-5 ) @require_flax class __UpperCamelCase ( a__ , unittest.TestCase ): def __a ( self ) -> List[Any]: a : str = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=lowerCAmelCase__ , text_from_pt=lowerCAmelCase__ , ) a : Any = 13 a : str = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) a : str = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) a : Optional[Any] = random_attention_mask([batch_size, 4] ) a : Optional[Any] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: a : Dict = FlaxViTModel(lowerCAmelCase__ ) a : Dict = FlaxBertModel(lowerCAmelCase__ ) return vision_model, text_model def __a ( self ) -> str: a : Union[str, Any] = FlaxViTModelTester(self ) a : Dict = FlaxBertModelTester(self ) a : str = vit_model_tester.prepare_config_and_inputs() a : Any = bert_model_tester.prepare_config_and_inputs() a, a : Optional[int] = vision_config_and_inputs a, a, a, a : Dict = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): def __a ( self ) -> List[Any]: a : str = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=lowerCAmelCase__ , text_from_pt=lowerCAmelCase__ , ) a : Tuple = 13 a : Union[str, Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) a : Union[str, Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) a : Tuple = random_attention_mask([batch_size, 4] ) a : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : List[Any] = FlaxCLIPVisionModel(lowerCAmelCase__ ) a : Tuple = FlaxBertModel(lowerCAmelCase__ ) return vision_model, text_model def __a ( self ) -> List[Any]: a : Tuple = FlaxCLIPVisionModelTester(self ) a : Union[str, Any] = FlaxBertModelTester(self ) a : Dict = clip_model_tester.prepare_config_and_inputs() a : Optional[int] = bert_model_tester.prepare_config_and_inputs() a, a : Dict = vision_config_and_inputs a, a, a, a : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Dict: a : str = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0 ) a : Optional[Any] = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) a : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) a : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ) a : Optional[Any] = model(**lowerCAmelCase__ ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) a : List[str] = np.array([[1.2_284_727, 0.3_104_122]] ) self.assertTrue(np.allclose(outputs.logits_per_image , lowerCAmelCase__ , atol=1E-3 ) )
79
0
def a_ ( _A , _A ) -> list: """simple docstring""" snake_case__ = word.split() def justify(_A , _A , _A ) -> str: snake_case__ = max_width - width snake_case__ = len(_A ) if len(_A ) == 1: # if there is only word in line # just insert overall_spaces_count for the remainder of line return line[0] + " " * overall_spaces_count else: snake_case__ = words_count - 1 # num_spaces_between_words_list[i] : tells you to insert # num_spaces_between_words_list[i] spaces # after word on line[i] snake_case__ = spaces_to_insert_between_words * [ overall_spaces_count // spaces_to_insert_between_words ] snake_case__ = ( overall_spaces_count % spaces_to_insert_between_words ) # distribute spaces via round robin to the left words for i in range(_A ): num_spaces_between_words_list[i] += 1 snake_case__ = [] for i in range(_A ): # add the word aligned_words_list.append(line[i] ) # add the spaces to insert aligned_words_list.append(num_spaces_between_words_list[i] * ' ' ) # just add the last word to the sentence aligned_words_list.append(line[-1] ) # join the aligned words list to form a justified line return "".join(_A ) snake_case__ = [] snake_case__ = [] snake_case__ = 0 for word in words: if width + len(_A ) + len(_A ) <= max_width: # keep adding words until we can fill out max_width # width = sum of length of all words (without overall_spaces_count) # len(word) = length of current word # len(line) = number of overall_spaces_count to insert between words line.append(_A ) width += len(_A ) else: # justify the line and add it to result answer.append(justify(_A , _A , _A ) ) # reset new line and new width snake_case__ , snake_case__ = [word], len(_A ) snake_case__ = max_width - width - len(_A ) answer.append(' '.join(_A ) + (remaining_spaces + 1) * ' ' ) return answer if __name__ == "__main__": from doctest import testmod testmod()
307
class __SCREAMING_SNAKE_CASE( a_ ): pass class __SCREAMING_SNAKE_CASE( a_ ): pass class __SCREAMING_SNAKE_CASE: def __init__( self: List[str] ) -> Union[str, Any]: snake_case__ = [ [], [], [], ] def lowerCAmelCase_ ( self: Union[str, Any] , UpperCamelCase: int , UpperCamelCase: int ) -> None: try: if len(self.queues[priority] ) >= 1_00: raise OverflowError('Maximum queue size is 100' ) self.queues[priority].append(UpperCamelCase ) except IndexError: raise ValueError('Valid priorities are 0, 1, and 2' ) def lowerCAmelCase_ ( self: List[Any] ) -> int: for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError('All queues are empty' ) def __str__( self: Union[str, Any] ) -> str: return "\n".join(F'''Priority {i}: {q}''' for i, q in enumerate(self.queues ) ) class __SCREAMING_SNAKE_CASE: def __init__( self: Union[str, Any] ) -> Any: snake_case__ = [] def lowerCAmelCase_ ( self: str , UpperCamelCase: int ) -> None: if len(self.queue ) == 1_00: raise OverFlowError('Maximum queue size is 100' ) self.queue.append(UpperCamelCase ) def lowerCAmelCase_ ( self: int ) -> int: if not self.queue: raise UnderFlowError('The queue is empty' ) else: snake_case__ = min(self.queue ) self.queue.remove(UpperCamelCase ) return data def __str__( self: Optional[Any] ) -> str: return str(self.queue ) def a_ ( ) -> List[Any]: """simple docstring""" snake_case__ = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_A ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_A ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def a_ ( ) -> List[Any]: """simple docstring""" snake_case__ = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_A ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_A ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
307
1
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase = logging.get_logger(__name__) def _lowerCamelCase( lowercase__ ) -> Dict: '''simple docstring''' __lowercase= torch.load(lowercase__ , map_location='cpu' ) if "model" in sd.keys(): __lowercase= torch.load(lowercase__ , map_location='cpu' )['model'] # pop unnecessary weights __lowercase= [ 'decoder.version', 'decoder.output_projection.weight', ] for key in keys_to_delete: if key in sd: sd.pop(lowercase__ ) __lowercase= { 'decoder.project_in_dim.weight': 'decoder.project_in.weight', 'decoder.project_out_dim.weight': 'decoder.project_out.weight', 'decoder.layer_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.layer_norm.bias': 'decoder.final_layer_norm.bias', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __lowercase= sd.pop(lowercase__ ) __lowercase= list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __lowercase= sd[key] # We split QKV in separate Q,K,V __lowercase= key.replace('.qkv_proj.' , '.q_proj.' ) __lowercase= key.replace('.qkv_proj.' , '.k_proj.' ) __lowercase= key.replace('.qkv_proj.' , '.v_proj.' ) __lowercase= value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __lowercase, __lowercase, __lowercase= torch.split(lowercase__ , depth // 3 , dim=0 ) __lowercase= q __lowercase= k __lowercase= v del sd[key] return sd @torch.no_grad() def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__=None ) -> Dict: '''simple docstring''' __lowercase= load_checkpoint(lowercase__ ) if config is not None: __lowercase= OPTConfig.from_pretrained(lowercase__ ) else: __lowercase= OPTConfig() __lowercase= OPTModel(lowercase__ ).half().eval() model.load_state_dict(lowercase__ ) # Check results Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) model.save_pretrained(lowercase__ ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') lowerCAmelCase = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
304
from __future__ import annotations from collections.abc import Callable def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__ , lowercase__ = 1_0_0 , ) -> float: '''simple docstring''' __lowercase= x_start __lowercase= fnc(lowercase__ ) __lowercase= 0.0 for _ in range(lowercase__ ): # Approximates small segments of curve as linear and solve # for trapezoidal area __lowercase= (x_end - x_start) / steps + xa __lowercase= fnc(lowercase__ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step __lowercase= xa __lowercase= fxa return area if __name__ == "__main__": def _lowerCamelCase( lowercase__ ) -> Dict: '''simple docstring''' return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') lowerCAmelCase = 1_0 while i <= 1_0_0_0_0_0: print(F'with {i} steps: {trapezoidal_area(f, -5, 5, i)}') i *= 1_0
304
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class lowerCamelCase : '''simple docstring''' def __init__( self: Optional[Any] , snake_case: Tuple , snake_case: Union[str, Any]=13 , snake_case: Optional[int]=7 , snake_case: Union[str, Any]=True , snake_case: Any=True , snake_case: str=True , snake_case: Dict=True , snake_case: int=99 , snake_case: Optional[Any]=32 , snake_case: Optional[Any]=2 , snake_case: Tuple=4 , snake_case: int=37 , snake_case: Any="gelu" , snake_case: List[str]=0.1 , snake_case: List[Any]=0.1 , snake_case: Optional[int]=512 , snake_case: int=16 , snake_case: str=2 , snake_case: int=0.0_2 , snake_case: str=False , snake_case: Any=True , snake_case: List[Any]="None" , snake_case: Optional[Any]=3 , snake_case: Optional[Any]=4 , snake_case: List[str]=None , ) -> str: snake_case_ :Optional[Any] = parent snake_case_ :str = batch_size snake_case_ :Tuple = seq_length snake_case_ :Any = is_training snake_case_ :Dict = use_input_mask snake_case_ :str = use_token_type_ids snake_case_ :List[str] = use_labels snake_case_ :Optional[int] = vocab_size snake_case_ :Dict = hidden_size snake_case_ :List[Any] = num_hidden_layers snake_case_ :Optional[Any] = num_attention_heads snake_case_ :int = intermediate_size snake_case_ :Optional[Any] = hidden_act snake_case_ :Dict = hidden_dropout_prob snake_case_ :str = attention_probs_dropout_prob snake_case_ :int = max_position_embeddings snake_case_ :int = type_vocab_size snake_case_ :Tuple = type_sequence_label_size snake_case_ :Dict = initializer_range snake_case_ :List[str] = num_labels snake_case_ :Union[str, Any] = num_choices snake_case_ :str = relative_attention snake_case_ :Any = position_biased_input snake_case_ :str = pos_att_type snake_case_ :int = scope def lowerCAmelCase_ ( self: List[str] ) -> Any: snake_case_ :str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ :List[Any] = None if self.use_input_mask: snake_case_ :Tuple = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ :List[str] = None if self.use_token_type_ids: snake_case_ :Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ :Union[str, Any] = None snake_case_ :Dict = None snake_case_ :List[str] = None if self.use_labels: snake_case_ :Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ :List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ :int = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=snake_case , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase_ ( self: Optional[int] , snake_case: List[str] , snake_case: Optional[int] , snake_case: str , snake_case: Any , snake_case: List[str] , snake_case: List[str] , snake_case: str ) -> Any: snake_case_ :Any = TFDebertaVaModel(config=snake_case ) snake_case_ :List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} snake_case_ :int = [input_ids, input_mask] snake_case_ :Dict = model(snake_case ) snake_case_ :List[Any] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self: Optional[Any] , snake_case: Optional[Any] , snake_case: List[Any] , snake_case: str , snake_case: int , snake_case: List[Any] , snake_case: int , snake_case: int ) -> Optional[Any]: snake_case_ :str = TFDebertaVaForMaskedLM(config=snake_case ) snake_case_ :Optional[Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } snake_case_ :Optional[Any] = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase_ ( self: int , snake_case: str , snake_case: List[Any] , snake_case: int , snake_case: Union[str, Any] , snake_case: Tuple , snake_case: int , snake_case: str ) -> str: snake_case_ :Tuple = self.num_labels snake_case_ :Tuple = TFDebertaVaForSequenceClassification(config=snake_case ) snake_case_ :List[str] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } snake_case_ :int = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase_ ( self: List[Any] , snake_case: Union[str, Any] , snake_case: int , snake_case: Tuple , snake_case: int , snake_case: Optional[int] , snake_case: str , snake_case: List[Any] ) -> List[str]: snake_case_ :Optional[Any] = self.num_labels snake_case_ :Any = TFDebertaVaForTokenClassification(config=snake_case ) snake_case_ :Union[str, Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } snake_case_ :int = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase_ ( self: List[str] , snake_case: Optional[int] , snake_case: str , snake_case: List[Any] , snake_case: Any , snake_case: Dict , snake_case: Union[str, Any] , snake_case: Optional[int] ) -> str: snake_case_ :Dict = TFDebertaVaForQuestionAnswering(config=snake_case ) snake_case_ :Any = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } snake_case_ :Tuple = model(snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase_ ( self: Union[str, Any] ) -> Any: snake_case_ :Optional[int] = self.prepare_config_and_inputs() ( ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ) :Tuple = config_and_inputs snake_case_ :List[str] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class lowerCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' _A : Dict = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) _A : Any = ( { """feature-extraction""": TFDebertaVaModel, """fill-mask""": TFDebertaVaForMaskedLM, """question-answering""": TFDebertaVaForQuestionAnswering, """text-classification""": TFDebertaVaForSequenceClassification, """token-classification""": TFDebertaVaForTokenClassification, """zero-shot""": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) _A : Any = False _A : List[Any] = False def lowerCAmelCase_ ( self: Dict ) -> List[str]: snake_case_ :Optional[Any] = TFDebertaVaModelTester(self ) snake_case_ :List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def lowerCAmelCase_ ( self: Union[str, Any] ) -> Any: self.config_tester.run_common_tests() def lowerCAmelCase_ ( self: Optional[Any] ) -> str: snake_case_ :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def lowerCAmelCase_ ( self: Optional[int] ) -> Optional[Any]: snake_case_ :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def lowerCAmelCase_ ( self: str ) -> List[Any]: snake_case_ :Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def lowerCAmelCase_ ( self: List[str] ) -> Optional[int]: snake_case_ :List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def lowerCAmelCase_ ( self: Any ) -> Optional[Any]: snake_case_ :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @slow def lowerCAmelCase_ ( self: Tuple ) -> List[Any]: snake_case_ :List[Any] = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(snake_case ) @require_tf class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @unittest.skip(reason="""Model not available yet""" ) def lowerCAmelCase_ ( self: Tuple ) -> int: pass @slow def lowerCAmelCase_ ( self: List[Any] ) -> int: snake_case_ :List[Any] = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) snake_case_ :Optional[int] = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) snake_case_ :int = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) snake_case_ :Optional[int] = model(snake_case , attention_mask=snake_case )[0] snake_case_ :Tuple = tf.constant( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , snake_case , atol=1E-4 )
66
"""simple docstring""" import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def A_ ( _lowercase, _lowercase ): '''simple docstring''' snake_case_ :int = XCLIPTextConfig() # derive patch size from model name snake_case_ :Union[str, Any] = model_name.find("""patch""" ) snake_case_ :List[str] = int(model_name[start_idx + len("""patch""" ) : start_idx + len("""patch""" ) + 2] ) snake_case_ :Any = XCLIPVisionConfig(patch_size=_lowercase, num_frames=_lowercase ) if "large" in model_name: snake_case_ :Optional[Any] = 768 snake_case_ :Union[str, Any] = 3072 snake_case_ :Any = 12 snake_case_ :Any = 1024 snake_case_ :str = 4096 snake_case_ :Union[str, Any] = 16 snake_case_ :Union[str, Any] = 24 snake_case_ :Tuple = 768 snake_case_ :Any = 3072 if model_name == "xclip-large-patch14-16-frames": snake_case_ :Any = 336 snake_case_ :Any = XCLIPConfig.from_text_vision_configs(_lowercase, _lowercase ) if "large" in model_name: snake_case_ :List[Any] = 768 return config def A_ ( _lowercase ): '''simple docstring''' if name == "token_embedding.weight": snake_case_ :Optional[Any] = name.replace("""token_embedding.weight""", """text_model.embeddings.token_embedding.weight""" ) if name == "positional_embedding": snake_case_ :Tuple = name.replace("""positional_embedding""", """text_model.embeddings.position_embedding.weight""" ) if "ln_1" in name: snake_case_ :Dict = name.replace("""ln_1""", """layer_norm1""" ) if "ln_2" in name: snake_case_ :str = name.replace("""ln_2""", """layer_norm2""" ) if "c_fc" in name: snake_case_ :str = name.replace("""c_fc""", """fc1""" ) if "c_proj" in name: snake_case_ :int = name.replace("""c_proj""", """fc2""" ) if name.startswith("""transformer.resblocks""" ): snake_case_ :Union[str, Any] = name.replace("""transformer.resblocks""", """text_model.encoder.layers""" ) if "attn.out_proj" in name and "message" not in name: snake_case_ :Union[str, Any] = name.replace("""attn.out_proj""", """self_attn.out_proj""" ) if "ln_final" in name: snake_case_ :Union[str, Any] = name.replace("""ln_final""", """text_model.final_layer_norm""" ) # visual encoder if name == "visual.class_embedding": snake_case_ :Any = name.replace("""visual.class_embedding""", """vision_model.embeddings.class_embedding""" ) if name == "visual.positional_embedding": snake_case_ :Optional[int] = name.replace("""visual.positional_embedding""", """vision_model.embeddings.position_embedding.weight""" ) if name.startswith("""visual.transformer.resblocks""" ): snake_case_ :Union[str, Any] = name.replace("""visual.transformer.resblocks""", """vision_model.encoder.layers""" ) if "visual.conv1" in name: snake_case_ :int = name.replace("""visual.conv1""", """vision_model.embeddings.patch_embedding""" ) if "visual.ln_pre" in name: snake_case_ :Any = name.replace("""visual.ln_pre""", """vision_model.pre_layernorm""" ) if "visual.ln_post" in name: snake_case_ :str = name.replace("""visual.ln_post""", """vision_model.post_layernorm""" ) if "visual.proj" in name: snake_case_ :Union[str, Any] = name.replace("""visual.proj""", """visual_projection.weight""" ) if "text_projection" in name: snake_case_ :Dict = name.replace("""text_projection""", """text_projection.weight""" ) # things on top if "prompts_visual_proj" in name: snake_case_ :List[str] = name.replace("""prompts_visual_proj""", """prompts_visual_projection""" ) if "prompts_visual_ln" in name: snake_case_ :Dict = name.replace("""prompts_visual_ln""", """prompts_visual_layernorm""" ) # mit if name == "mit.positional_embedding": snake_case_ :str = name.replace("""positional""", """position""" ) if name.startswith("""mit.resblocks""" ): snake_case_ :Dict = name.replace("""mit.resblocks""", """mit.encoder.layers""" ) # prompts generator if name.startswith("""prompts_generator.norm""" ): snake_case_ :Union[str, Any] = name.replace("""prompts_generator.norm""", """prompts_generator.layernorm""" ) return name def A_ ( _lowercase, _lowercase ): '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ :Dict = orig_state_dict.pop(_lowercase ) if "attn.in_proj" in key: snake_case_ :Optional[Any] = key.split(""".""" ) if key.startswith("""visual""" ): snake_case_ :Any = key_split[3] snake_case_ :Optional[Any] = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: snake_case_ :str = val[ :dim, : ] snake_case_ :Optional[int] = val[ dim : dim * 2, : ] snake_case_ :Union[str, Any] = val[ -dim:, : ] else: snake_case_ :Dict = val[ :dim ] snake_case_ :Optional[int] = val[ dim : dim * 2 ] snake_case_ :Optional[int] = val[ -dim: ] else: if "weight" in key: snake_case_ :Optional[Any] = val[ :dim, : ] snake_case_ :List[str] = val[ dim : dim * 2, : ] snake_case_ :Dict = val[ -dim:, : ] else: snake_case_ :Union[str, Any] = val[:dim] snake_case_ :Union[str, Any] = val[ dim : dim * 2 ] snake_case_ :Union[str, Any] = val[-dim:] elif key.startswith("""mit""" ): snake_case_ :Tuple = key_split[2] snake_case_ :Union[str, Any] = config.vision_config.mit_hidden_size if "weight" in key: snake_case_ :Optional[int] = val[:dim, :] snake_case_ :Optional[int] = val[dim : dim * 2, :] snake_case_ :str = val[-dim:, :] else: snake_case_ :str = val[:dim] snake_case_ :Any = val[dim : dim * 2] snake_case_ :int = val[-dim:] else: snake_case_ :Tuple = key_split[2] snake_case_ :Any = config.text_config.hidden_size if "weight" in key: snake_case_ :Dict = val[:dim, :] snake_case_ :Dict = val[ dim : dim * 2, : ] snake_case_ :List[str] = val[-dim:, :] else: snake_case_ :Any = val[:dim] snake_case_ :Tuple = val[ dim : dim * 2 ] snake_case_ :List[str] = val[-dim:] else: snake_case_ :Optional[int] = rename_key(_lowercase ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: snake_case_ :Optional[Any] = val.T snake_case_ :Tuple = val return orig_state_dict def A_ ( _lowercase ): '''simple docstring''' if num_frames == 8: snake_case_ :str = """eating_spaghetti_8_frames.npy""" elif num_frames == 16: snake_case_ :int = """eating_spaghetti.npy""" elif num_frames == 32: snake_case_ :List[str] = """eating_spaghetti_32_frames.npy""" snake_case_ :int = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""", filename=_lowercase, repo_type="""dataset""", ) snake_case_ :Union[str, Any] = np.load(_lowercase ) return list(_lowercase ) def A_ ( _lowercase, _lowercase=None, _lowercase=False ): '''simple docstring''' snake_case_ :List[Any] = { # fully supervised kinetics-400 checkpoints """xclip-base-patch32""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth""", """xclip-base-patch32-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth""" ), """xclip-base-patch16""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth""", """xclip-base-patch16-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth""" ), """xclip-large-patch14""": """https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb""", """xclip-large-patch14-16-frames""": """https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f""", # fully supervised kinetics-600 checkpoints """xclip-base-patch16-kinetics-600""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth""" ), """xclip-base-patch16-kinetics-600-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth""" ), """xclip-large-patch14-kinetics-600""": """https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be""", # few shot """xclip-base-patch16-hmdb-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth""" ), """xclip-base-patch16-hmdb-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth""" ), """xclip-base-patch16-hmdb-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth""" ), """xclip-base-patch16-hmdb-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth""" ), """xclip-base-patch16-ucf-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth""" ), """xclip-base-patch16-ucf-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth""" ), """xclip-base-patch16-ucf-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth""" ), """xclip-base-patch16-ucf-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth""" ), # zero shot """xclip-base-patch16-zero-shot""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth""", } snake_case_ :Optional[int] = model_to_url[model_name] snake_case_ :int = 8 if "16-frames" in model_name: snake_case_ :List[Any] = 16 elif "shot" in model_name: snake_case_ :Dict = 32 snake_case_ :Optional[int] = get_xclip_config(_lowercase, _lowercase ) snake_case_ :Optional[Any] = XCLIPModel(_lowercase ) model.eval() if "drive" in checkpoint_url: snake_case_ :List[str] = """pytorch_model.bin""" gdown.cached_download(_lowercase, _lowercase, quiet=_lowercase ) snake_case_ :List[Any] = torch.load(_lowercase, map_location="""cpu""" )["""model"""] else: snake_case_ :Tuple = torch.hub.load_state_dict_from_url(_lowercase )["""model"""] snake_case_ :Union[str, Any] = convert_state_dict(_lowercase, _lowercase ) snake_case_ :str = XCLIPModel(_lowercase ) snake_case_, snake_case_ :Optional[int] = model.load_state_dict(_lowercase, strict=_lowercase ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() snake_case_ :List[str] = 336 if model_name == """xclip-large-patch14-16-frames""" else 224 snake_case_ :List[Any] = VideoMAEImageProcessor(size=_lowercase ) snake_case_ :Any = CLIPTokenizer.from_pretrained("""openai/clip-vit-base-patch32""" ) snake_case_ :str = CLIPTokenizerFast.from_pretrained("""openai/clip-vit-base-patch32""" ) snake_case_ :Optional[Any] = XCLIPProcessor(image_processor=_lowercase, tokenizer=_lowercase ) snake_case_ :Optional[int] = prepare_video(_lowercase ) snake_case_ :Optional[Any] = processor( text=["""playing sports""", """eating spaghetti""", """go shopping"""], videos=_lowercase, return_tensors="""pt""", padding=_lowercase ) print("""Shape of pixel values:""", inputs.pixel_values.shape ) with torch.no_grad(): snake_case_ :List[Any] = model(**_lowercase ) # Verify outputs snake_case_ :List[Any] = outputs.logits_per_video snake_case_ :Any = logits_per_video.softmax(dim=1 ) print("""Probs:""", _lowercase ) # kinetics-400 if model_name == "xclip-base-patch32": snake_case_ :Union[str, Any] = torch.tensor([[0.0019, 0.9951, 0.0030]] ) elif model_name == "xclip-base-patch32-16-frames": snake_case_ :str = torch.tensor([[7.09_99e-04, 9.98_83e-01, 4.55_80e-04]] ) elif model_name == "xclip-base-patch16": snake_case_ :Tuple = torch.tensor([[0.0083, 0.9681, 0.0236]] ) elif model_name == "xclip-base-patch16-16-frames": snake_case_ :Any = torch.tensor([[7.69_37e-04, 9.97_28e-01, 1.94_73e-03]] ) elif model_name == "xclip-large-patch14": snake_case_ :str = torch.tensor([[0.0062, 0.9864, 0.0075]] ) elif model_name == "xclip-large-patch14-16-frames": snake_case_ :Tuple = torch.tensor([[3.38_77e-04, 9.99_37e-01, 2.88_88e-04]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": snake_case_ :List[Any] = torch.tensor([[0.0555, 0.8914, 0.0531]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": snake_case_ :Union[str, Any] = torch.tensor([[3.85_54e-04, 9.99_29e-01, 3.27_54e-04]] ) elif model_name == "xclip-large-patch14-kinetics-600": snake_case_ :List[Any] = torch.tensor([[0.0036, 0.9920, 0.0045]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": snake_case_ :Dict = torch.tensor([[7.18_90e-06, 9.99_94e-01, 5.65_59e-05]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": snake_case_ :Union[str, Any] = torch.tensor([[1.03_20e-05, 9.99_93e-01, 6.24_35e-05]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": snake_case_ :str = torch.tensor([[4.13_77e-06, 9.99_90e-01, 9.83_86e-05]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": snake_case_ :str = torch.tensor([[4.13_47e-05, 9.99_62e-01, 3.34_11e-04]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": snake_case_ :int = torch.tensor([[8.58_57e-05, 9.99_28e-01, 6.32_91e-04]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": snake_case_ :Optional[int] = torch.tensor([[8.58_57e-05, 9.99_28e-01, 6.32_91e-04]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": snake_case_ :Any = torch.tensor([[0.0027, 0.9904, 0.0070]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": snake_case_ :Tuple = torch.tensor([[9.82_19e-04, 9.95_93e-01, 3.08_63e-03]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": snake_case_ :Union[str, Any] = torch.tensor([[3.50_82e-04, 9.97_85e-01, 1.79_66e-03]] ) else: raise ValueError(f"""Model name {model_name} not supported""" ) assert torch.allclose(_lowercase, _lowercase, atol=1e-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowercase ) if push_to_hub: print("""Pushing model, processor and slow tokenizer files to the hub...""" ) model.push_to_hub(_lowercase, organization="""nielsr""" ) processor.push_to_hub(_lowercase, organization="""nielsr""" ) slow_tokenizer.push_to_hub(_lowercase, organization="""nielsr""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="xclip-base-patch32", type=str, help="Name of the model.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __a = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any]) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase : List[str] = {} __UpperCamelCase : List[Any] = tokenizer(example["content"] , truncation=_lowerCamelCase)["input_ids"] __UpperCamelCase : Union[str, Any] = len(example["content"]) / len(output["input_ids"]) return output lowercase : str = HfArgumentParser(PretokenizationArguments) lowercase : Optional[Any] = parser.parse_args() if args.num_workers is None: lowercase : Optional[int] = multiprocessing.cpu_count() lowercase : Any = AutoTokenizer.from_pretrained(args.tokenizer_dir) lowercase : Any = time.time() lowercase : str = load_dataset(args.dataset_name, split='train') print(f"Dataset loaded in {time.time()-t_start:.2f}s") lowercase : Optional[Any] = time.time() lowercase : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ 'repo_name', 'path', 'copies', 'size', 'content', 'license', 'hash', 'line_mean', 'line_max', 'alpha_frac', 'autogenerated', ], ) print(f"Dataset tokenized in {time.time()-t_start:.2f}s") lowercase : str = time.time() ds.push_to_hub(args.tokenized_data_repo) print(f"Data pushed to the hub in {time.time()-t_start:.2f}s")
151
from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class lowerCamelCase__ : '''simple docstring''' _A = 42 _A = 42 class lowerCamelCase__ : '''simple docstring''' def __init__( self :Optional[Any] , a :int ) -> Tuple: __UpperCamelCase : list[list[Edge]] = [[] for _ in range(a )] __UpperCamelCase : str = size def __getitem__( self :str , a :int ) -> Iterator[Edge]: return iter(self._graph[vertex] ) @property def _lowerCamelCase ( self :Any ) -> List[str]: return self._size def _lowerCamelCase ( self :Dict , a :int , a :int , a :int ) -> Any: if weight not in (0, 1): raise ValueError("Edge weight must be either 0 or 1." ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("Vertex indexes must be in [0; size)." ) self._graph[from_vertex].append(Edge(a , a ) ) def _lowerCamelCase ( self :List[str] , a :int , a :int ) -> int | None: __UpperCamelCase : Union[str, Any] = deque([start_vertex] ) __UpperCamelCase : list[int | None] = [None] * self.size __UpperCamelCase : Dict = 0 while queue: __UpperCamelCase : Tuple = queue.popleft() __UpperCamelCase : int = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: __UpperCamelCase : Optional[Any] = current_distance + edge.weight __UpperCamelCase : Dict = distances[edge.destination_vertex] if ( isinstance(a , a ) and new_distance >= dest_vertex_distance ): continue __UpperCamelCase : Optional[Any] = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("No path from start_vertex to finish_vertex." ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
151
1
'''simple docstring''' import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging __snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name class lowercase ( A__ ): """simple docstring""" def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , ): '''simple docstring''' super().__init__() if safety_checker is None: logger.warning( F'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ''' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered''' ''' results in services or applications open to the public. Both the diffusers team and Hugging Face''' ''' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling''' ''' it only for use-cases that involve analyzing network behavior or auditing its results. For more''' ''' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .''' ) self.register_modules( speech_model=UpperCamelCase_ , speech_processor=UpperCamelCase_ , vae=UpperCamelCase_ , text_encoder=UpperCamelCase_ , tokenizer=UpperCamelCase_ , unet=UpperCamelCase_ , scheduler=UpperCamelCase_ , feature_extractor=UpperCamelCase_ , ) def lowerCAmelCase__ ( self , UpperCamelCase_ = "auto" ): '''simple docstring''' if slice_size == "auto": UpperCamelCase__ :int = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.enable_attention_slicing(UpperCamelCase_ ) @torch.no_grad() def __call__( self , UpperCamelCase_ , UpperCamelCase_=16000 , UpperCamelCase_ = 512 , UpperCamelCase_ = 512 , UpperCamelCase_ = 50 , UpperCamelCase_ = 7.5 , UpperCamelCase_ = None , UpperCamelCase_ = 1 , UpperCamelCase_ = 0.0 , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = "pil" , UpperCamelCase_ = True , UpperCamelCase_ = None , UpperCamelCase_ = 1 , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :int = self.speech_processor.feature_extractor( UpperCamelCase_ , return_tensors='''pt''' , sampling_rate=UpperCamelCase_ ).input_features.to(self.device ) UpperCamelCase__ :List[str] = self.speech_model.generate(UpperCamelCase_ , max_length=480000 ) UpperCamelCase__ :Optional[Any] = self.speech_processor.tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ , normalize=UpperCamelCase_ )[ 0 ] if isinstance(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Dict = 1 elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Optional[Any] = len(UpperCamelCase_ ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(UpperCamelCase_ )}''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(UpperCamelCase_ , UpperCamelCase_ ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(UpperCamelCase_ )}.''' ) # get prompt text embeddings UpperCamelCase__ :Optional[int] = self.tokenizer( UpperCamelCase_ , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) UpperCamelCase__ :List[Any] = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: UpperCamelCase__ :Optional[int] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) UpperCamelCase__ :Tuple = text_input_ids[:, : self.tokenizer.model_max_length] UpperCamelCase__ :Dict = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ :Any = text_embeddings.shape UpperCamelCase__ :Dict = text_embeddings.repeat(1 , UpperCamelCase_ , 1 ) UpperCamelCase__ :Dict = text_embeddings.view(bs_embed * num_images_per_prompt , UpperCamelCase_ , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. UpperCamelCase__ :Dict = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: UpperCamelCase__ :List[str] if negative_prompt is None: UpperCamelCase__ :Dict = [''''''] * batch_size elif type(UpperCamelCase_ ) is not type(UpperCamelCase_ ): raise TypeError( F'''`negative_prompt` should be the same type to `prompt`, but got {type(UpperCamelCase_ )} !=''' F''' {type(UpperCamelCase_ )}.''' ) elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Tuple = [negative_prompt] elif batch_size != len(UpperCamelCase_ ): raise ValueError( F'''`negative_prompt`: {negative_prompt} has batch size {len(UpperCamelCase_ )}, but `prompt`:''' F''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' ''' the batch size of `prompt`.''' ) else: UpperCamelCase__ :int = negative_prompt UpperCamelCase__ :Optional[Any] = text_input_ids.shape[-1] UpperCamelCase__ :List[Any] = self.tokenizer( UpperCamelCase_ , padding='''max_length''' , max_length=UpperCamelCase_ , truncation=UpperCamelCase_ , return_tensors='''pt''' , ) UpperCamelCase__ :Optional[int] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method UpperCamelCase__ :Tuple = uncond_embeddings.shape[1] UpperCamelCase__ :Any = uncond_embeddings.repeat(1 , UpperCamelCase_ , 1 ) UpperCamelCase__ :List[Any] = uncond_embeddings.view(batch_size * num_images_per_prompt , UpperCamelCase_ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCamelCase__ :Tuple = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. UpperCamelCase__ :Dict = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) UpperCamelCase__ :int = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps UpperCamelCase__ :Optional[int] = torch.randn(UpperCamelCase_ , generator=UpperCamelCase_ , device='''cpu''' , dtype=UpperCamelCase_ ).to( self.device ) else: UpperCamelCase__ :Tuple = torch.randn(UpperCamelCase_ , generator=UpperCamelCase_ , device=self.device , dtype=UpperCamelCase_ ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) UpperCamelCase__ :str = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(UpperCamelCase_ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand UpperCamelCase__ :List[Any] = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler UpperCamelCase__ :List[str] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCamelCase__ :Tuple = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCamelCase__ :List[Any] = {} if accepts_eta: UpperCamelCase__ :List[str] = eta for i, t in enumerate(self.progress_bar(UpperCamelCase_ ) ): # expand the latents if we are doing classifier free guidance UpperCamelCase__ :int = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCamelCase__ :int = self.scheduler.scale_model_input(UpperCamelCase_ , UpperCamelCase_ ) # predict the noise residual UpperCamelCase__ :List[Any] = self.unet(UpperCamelCase_ , UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ ).sample # perform guidance if do_classifier_free_guidance: UpperCamelCase__ , UpperCamelCase__ :List[Any] = noise_pred.chunk(2 ) UpperCamelCase__ :int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase__ :Tuple = self.scheduler.step(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase__ :Dict = 1 / 0.18215 * latents UpperCamelCase__ :Dict = self.vae.decode(UpperCamelCase_ ).sample UpperCamelCase__ :Dict = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 UpperCamelCase__ :List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCamelCase__ :Optional[int] = self.numpy_to_pil(UpperCamelCase_ ) if not return_dict: return image return StableDiffusionPipelineOutput(images=UpperCamelCase_ , nsfw_content_detected=UpperCamelCase_ )
97
def lowercase_ ( _lowerCamelCase : int): lowercase__ : Dict = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
87
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { '''microsoft/git-base''': '''https://huggingface.co/microsoft/git-base/resolve/main/config.json''', } class __a ( __UpperCamelCase ): __lowercase : Tuple = 'git_vision_model' def __init__( self , lowerCAmelCase__=768 , lowerCAmelCase__=3_072 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=3 , lowerCAmelCase__=224 , lowerCAmelCase__=16 , lowerCAmelCase__="quick_gelu" , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0_2 , **lowerCAmelCase__ , ) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowercase__: Tuple = hidden_size lowercase__: Optional[Any] = intermediate_size lowercase__: Any = num_hidden_layers lowercase__: Any = num_attention_heads lowercase__: Union[str, Any] = num_channels lowercase__: Optional[int] = patch_size lowercase__: Tuple = image_size lowercase__: Union[str, Any] = initializer_range lowercase__: Dict = attention_dropout lowercase__: Dict = layer_norm_eps lowercase__: Optional[int] = hidden_act @classmethod def SCREAMING_SNAKE_CASE__ ( cls , lowerCAmelCase__ , **lowerCAmelCase__ ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(lowerCAmelCase__ ) lowercase__ , lowercase__: List[Any] = cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get('model_type' ) == "git": lowercase__: int = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class __a ( __UpperCamelCase ): __lowercase : Tuple = 'git' def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=30_522 , lowerCAmelCase__=768 , lowerCAmelCase__=6 , lowerCAmelCase__=12 , lowerCAmelCase__=3_072 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_024 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__="absolute" , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=101 , lowerCAmelCase__=102 , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> List[str]: '''simple docstring''' super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) if vision_config is None: lowercase__: str = {} logger.info('vision_config is None. initializing the GitVisionConfig with default values.' ) lowercase__: str = GitVisionConfig(**lowerCAmelCase__ ) lowercase__: Any = vocab_size lowercase__: str = hidden_size lowercase__: Any = num_hidden_layers lowercase__: List[str] = num_attention_heads lowercase__: Dict = hidden_act lowercase__: List[Any] = intermediate_size lowercase__: Tuple = hidden_dropout_prob lowercase__: Tuple = attention_probs_dropout_prob lowercase__: Tuple = max_position_embeddings lowercase__: str = initializer_range lowercase__: List[str] = layer_norm_eps lowercase__: Union[str, Any] = position_embedding_type lowercase__: Tuple = use_cache lowercase__: int = tie_word_embeddings lowercase__: Optional[int] = num_image_with_embedding lowercase__: Optional[Any] = bos_token_id lowercase__: Optional[Any] = eos_token_id def SCREAMING_SNAKE_CASE__ ( self ) -> List[Any]: '''simple docstring''' lowercase__: Tuple = copy.deepcopy(self.__dict__ ) lowercase__: List[Any] = self.vision_config.to_dict() lowercase__: List[str] = self.__class__.model_type return output
288
def snake_case_ ( snake_case ) -> list[int]: lowercase__: Dict = [0 for i in range(len(snake_case ) )] # initialize interval's left pointer and right pointer lowercase__ , lowercase__: Union[str, Any] = 0, 0 for i in range(1 , len(snake_case ) ): # case when current index is inside the interval if i <= right_pointer: lowercase__: List[Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) lowercase__: List[str] = min_edge while go_next(snake_case , snake_case , snake_case ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: lowercase__ , lowercase__: List[Any] = i, i + z_result[i] - 1 return z_result def snake_case_ ( snake_case , snake_case , snake_case ) -> bool: return i + z_result[i] < len(snake_case ) and s[z_result[i]] == s[i + z_result[i]] def snake_case_ ( snake_case , snake_case ) -> int: lowercase__: Tuple = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string lowercase__: Any = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(snake_case ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
288
1
'''simple docstring''' import pickle import numpy as np from matplotlib import pyplot as plt class _a : def __init__( self : Optional[int] , lowercase : List[Any] , lowercase : Any , lowercase : Optional[Any] , lowercase : Optional[int] , lowercase : Union[str, Any] , lowercase : int=0.2 , lowercase : Optional[int]=0.2 ): '''simple docstring''' UpperCAmelCase = bp_numa UpperCAmelCase = bp_numa UpperCAmelCase = bp_numa UpperCAmelCase = conva_get[:2] UpperCAmelCase = conva_get[2] UpperCAmelCase = size_pa UpperCAmelCase = rate_w UpperCAmelCase = rate_t UpperCAmelCase = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCAmelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCAmelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCAmelCase = -2 * np.random.rand(self.conva[1] ) + 1 UpperCAmelCase = -2 * np.random.rand(self.num_bpa ) + 1 UpperCAmelCase = -2 * np.random.rand(self.num_bpa ) + 1 def A ( self : List[str] , lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = { '''num_bp1''': self.num_bpa, '''num_bp2''': self.num_bpa, '''num_bp3''': self.num_bpa, '''conv1''': self.conva, '''step_conv1''': self.step_conva, '''size_pooling1''': self.size_poolinga, '''rate_weight''': self.rate_weight, '''rate_thre''': self.rate_thre, '''w_conv1''': self.w_conva, '''wkj''': self.wkj, '''vji''': self.vji, '''thre_conv1''': self.thre_conva, '''thre_bp2''': self.thre_bpa, '''thre_bp3''': self.thre_bpa, } with open(lowercase , '''wb''' ) as f: pickle.dump(lowercase , lowercase ) print(f"Model saved: {save_path}" ) @classmethod def A ( cls : Union[str, Any] , lowercase : Union[str, Any] ): '''simple docstring''' with open(lowercase , '''rb''' ) as f: UpperCAmelCase = pickle.load(lowercase ) # noqa: S301 UpperCAmelCase = model_dic.get('''conv1''' ) conv_get.append(model_dic.get('''step_conv1''' ) ) UpperCAmelCase = model_dic.get('''size_pooling1''' ) UpperCAmelCase = model_dic.get('''num_bp1''' ) UpperCAmelCase = model_dic.get('''num_bp2''' ) UpperCAmelCase = model_dic.get('''num_bp3''' ) UpperCAmelCase = model_dic.get('''rate_weight''' ) UpperCAmelCase = model_dic.get('''rate_thre''' ) # create model instance UpperCAmelCase = CNN(lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) # modify model parameter UpperCAmelCase = model_dic.get('''w_conv1''' ) UpperCAmelCase = model_dic.get('''wkj''' ) UpperCAmelCase = model_dic.get('''vji''' ) UpperCAmelCase = model_dic.get('''thre_conv1''' ) UpperCAmelCase = model_dic.get('''thre_bp2''' ) UpperCAmelCase = model_dic.get('''thre_bp3''' ) return conv_ins def A ( self : Dict , lowercase : Union[str, Any] ): '''simple docstring''' return 1 / (1 + np.exp(-1 * x )) def A ( self : List[str] , lowercase : str ): '''simple docstring''' return round(lowercase , 3 ) def A ( self : Tuple , lowercase : Dict , lowercase : Any , lowercase : Tuple , lowercase : Dict , lowercase : Tuple ): '''simple docstring''' UpperCAmelCase = convs[0] UpperCAmelCase = convs[1] UpperCAmelCase = np.shape(lowercase )[0] # get the data slice of original image data, data_focus UpperCAmelCase = [] for i_focus in range(0 , size_data - size_conv + 1 , lowercase ): for j_focus in range(0 , size_data - size_conv + 1 , lowercase ): UpperCAmelCase = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCAmelCase = [] UpperCAmelCase = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(lowercase ): UpperCAmelCase = [] for i_focus in range(len(lowercase ) ): UpperCAmelCase = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(lowercase ) ) UpperCAmelCase = np.asmatrix(lowercase ).reshape( lowercase , lowercase ) data_featuremap.append(lowercase ) # expanding the data slice to One dimenssion UpperCAmelCase = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(lowercase ) ) UpperCAmelCase = np.asarray(lowercase ) return focus_list, data_featuremap def A ( self : Any , lowercase : str , lowercase : Optional[int] , lowercase : List[Any]="average_pool" ): '''simple docstring''' UpperCAmelCase = len(featuremaps[0] ) UpperCAmelCase = int(size_map / size_pooling ) UpperCAmelCase = [] for i_map in range(len(lowercase ) ): UpperCAmelCase = featuremaps[i_map] UpperCAmelCase = [] for i_focus in range(0 , lowercase , lowercase ): for j_focus in range(0 , lowercase , lowercase ): UpperCAmelCase = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(lowercase ) ) UpperCAmelCase = np.asmatrix(lowercase ).reshape(lowercase , lowercase ) featuremap_pooled.append(lowercase ) return featuremap_pooled def A ( self : Union[str, Any] , lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = [] for i in range(len(lowercase ) ): UpperCAmelCase = np.shape(data[i] ) UpperCAmelCase = data[i].reshape(1 , shapes[0] * shapes[1] ) UpperCAmelCase = data_listed.getA().tolist()[0] data_expanded.extend(lowercase ) UpperCAmelCase = np.asarray(lowercase ) return data_expanded def A ( self : int , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = np.asarray(lowercase ) UpperCAmelCase = np.shape(lowercase ) UpperCAmelCase = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def A ( self : Any , lowercase : int , lowercase : List[Any] , lowercase : int , lowercase : Any , lowercase : str ): '''simple docstring''' UpperCAmelCase = [] UpperCAmelCase = 0 for i_map in range(lowercase ): UpperCAmelCase = np.ones((size_map, size_map) ) for i in range(0 , lowercase , lowercase ): for j in range(0 , lowercase , lowercase ): UpperCAmelCase = pd_pool[ i_pool ] UpperCAmelCase = i_pool + 1 UpperCAmelCase = np.multiply( lowercase , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(lowercase ) return pd_all def A ( self : Optional[Any] , lowercase : str , lowercase : Union[str, Any] , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : List[Any] , lowercase : str=bool ): '''simple docstring''' print('''----------------------Start Training-------------------------''' ) print((''' - - Shape: Train_Data ''', np.shape(lowercase )) ) print((''' - - Shape: Teach_Data ''', np.shape(lowercase )) ) UpperCAmelCase = 0 UpperCAmelCase = [] UpperCAmelCase = 10_000 while rp < n_repeat and mse >= error_accuracy: UpperCAmelCase = 0 print(f"-------------Learning Time {rp}--------------" ) for p in range(len(lowercase ) ): # print('------------Learning Image: %d--------------'%p) UpperCAmelCase = np.asmatrix(datas_train[p] ) UpperCAmelCase = np.asarray(datas_teach[p] ) UpperCAmelCase , UpperCAmelCase = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase = self.pooling(lowercase , self.size_poolinga ) UpperCAmelCase = np.shape(lowercase ) UpperCAmelCase = self._expand(lowercase ) UpperCAmelCase = data_bp_input UpperCAmelCase = np.dot(lowercase , self.vji.T ) - self.thre_bpa UpperCAmelCase = self.sig(lowercase ) UpperCAmelCase = np.dot(lowercase , self.wkj.T ) - self.thre_bpa UpperCAmelCase = self.sig(lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCAmelCase = np.multiply( (data_teach - bp_outa) , np.multiply(lowercase , (1 - bp_outa) ) ) UpperCAmelCase = np.multiply( np.dot(lowercase , self.wkj ) , np.multiply(lowercase , (1 - bp_outa) ) ) UpperCAmelCase = np.dot(lowercase , self.vji ) UpperCAmelCase = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCAmelCase = pd_conva_pooled.T.getA().tolist() UpperCAmelCase = self._calculate_gradient_from_pool( lowercase , lowercase , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCAmelCase = self._expand_mat(pd_conva_all[k_conv] ) UpperCAmelCase = self.rate_weight * np.dot(lowercase , lowercase ) UpperCAmelCase = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCAmelCase = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCAmelCase = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCAmelCase = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCAmelCase = self.thre_bpa - pd_k_all * self.rate_thre UpperCAmelCase = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCAmelCase = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCAmelCase = rp + 1 UpperCAmelCase = error_count / patterns all_mse.append(lowercase ) def draw_error(): UpperCAmelCase = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(lowercase , '''+-''' ) plt.plot(lowercase , '''r--''' ) plt.xlabel('''Learning Times''' ) plt.ylabel('''All_mse''' ) plt.grid(lowercase , alpha=0.5 ) plt.show() print('''------------------Training Complished---------------------''' ) print((''' - - Training epoch: ''', rp, f" - - Mse: {mse:.6f}") ) if draw_e: draw_error() return mse def A ( self : List[str] , lowercase : Optional[Any] ): '''simple docstring''' UpperCAmelCase = [] print('''-------------------Start Testing-------------------------''' ) print((''' - - Shape: Test_Data ''', np.shape(lowercase )) ) for p in range(len(lowercase ) ): UpperCAmelCase = np.asmatrix(datas_test[p] ) UpperCAmelCase , UpperCAmelCase = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase = self.pooling(lowercase , self.size_poolinga ) UpperCAmelCase = self._expand(lowercase ) UpperCAmelCase = data_bp_input UpperCAmelCase = bp_outa * self.vji.T - self.thre_bpa UpperCAmelCase = self.sig(lowercase ) UpperCAmelCase = bp_outa * self.wkj.T - self.thre_bpa UpperCAmelCase = self.sig(lowercase ) produce_out.extend(bp_outa.getA().tolist() ) UpperCAmelCase = [list(map(self.do_round , lowercase ) ) for each in produce_out] return np.asarray(lowercase ) def A ( self : Tuple , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = np.asmatrix(lowercase ) UpperCAmelCase , UpperCAmelCase = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase = self.pooling(lowercase , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
34
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : Union[str, Any] = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowerCAmelCase (lowercase_ ): '''simple docstring''' lowerCAmelCase__ : List[str] = """megatron-bert""" def __init__(self : Tuple , UpperCamelCase : Optional[int]=29056 , UpperCamelCase : Optional[Any]=1024 , UpperCamelCase : Any=24 , UpperCamelCase : int=16 , UpperCamelCase : Optional[int]=4096 , UpperCamelCase : int="gelu" , UpperCamelCase : int=0.1 , UpperCamelCase : Tuple=0.1 , UpperCamelCase : Any=512 , UpperCamelCase : int=2 , UpperCamelCase : Dict=0.02 , UpperCamelCase : Dict=1E-12 , UpperCamelCase : List[Any]=0 , UpperCamelCase : Optional[int]="absolute" , UpperCamelCase : List[Any]=True , **UpperCamelCase : str , ): '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase , **UpperCamelCase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache
2
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: UpperCAmelCase_ = None UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} UpperCAmelCase_ = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } UpperCAmelCase_ = { 'albert-base-v1': 5_1_2, 'albert-large-v1': 5_1_2, 'albert-xlarge-v1': 5_1_2, 'albert-xxlarge-v1': 5_1_2, 'albert-base-v2': 5_1_2, 'albert-large-v2': 5_1_2, 'albert-xlarge-v2': 5_1_2, 'albert-xxlarge-v2': 5_1_2, } UpperCAmelCase_ = '▁' class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = VOCAB_FILES_NAMES lowerCAmelCase_ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Tuple = AlbertTokenizer def __init__( self : Optional[Any] , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : int=True , _UpperCAmelCase : Tuple=True , _UpperCAmelCase : List[str]=False , _UpperCAmelCase : List[str]="[CLS]" , _UpperCAmelCase : List[Any]="[SEP]" , _UpperCAmelCase : int="<unk>" , _UpperCAmelCase : int="[SEP]" , _UpperCAmelCase : List[str]="<pad>" , _UpperCAmelCase : List[Any]="[CLS]" , _UpperCAmelCase : Union[str, Any]="[MASK]" , **_UpperCAmelCase : Dict , ): """simple docstring""" UpperCAmelCase__ = ( AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase , normalized=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else mask_token ) super().__init__( _UpperCAmelCase , tokenizer_file=_UpperCAmelCase , do_lower_case=_UpperCAmelCase , remove_space=_UpperCAmelCase , keep_accents=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , **_UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case UpperCAmelCase__ = remove_space UpperCAmelCase__ = keep_accents UpperCAmelCase__ = vocab_file UpperCAmelCase__ = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE__ ( self : int , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(_UpperCAmelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( _UpperCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ): copyfile(self.vocab_file , _UpperCAmelCase ) return (out_vocab_file,)
61
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' while b: UpperCAmelCase__ , UpperCAmelCase__ = b, a % b return a def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(SCREAMING_SNAKE_CASE__ , a % b ) def _UpperCamelCase ( ): '''simple docstring''' print(F'''euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}''' ) print(F'''euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}''' ) print(F'''euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}''' ) print(F'''euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}''' ) print(F'''euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}''' ) print(F'''euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}''' ) print(F'''euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}''' ) print(F'''euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}''' ) if __name__ == "__main__": main()
61
1
'''simple docstring''' def _lowerCamelCase ( lowercase : float , lowercase : float ) -> float: if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
63
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class __a (unittest.TestCase): '''simple docstring''' def _a ( self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ : Any = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """external_deps""", """test_metrics.py"""] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 SCREAMING_SNAKE_CASE__ : Optional[int] = test_metrics @require_cpu def _a ( self ) -> List[Any]: """simple docstring""" debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def _a ( self ) -> List[str]: """simple docstring""" debug_launcher(self.test_metrics.main ) @require_single_gpu def _a ( self ) -> int: """simple docstring""" self.test_metrics.main() @require_multi_gpu def _a ( self ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ : List[Any] = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_a , env=os.environ.copy() )
132
0
"""simple docstring""" import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter _SCREAMING_SNAKE_CASE : str = True except ImportError: _SCREAMING_SNAKE_CASE : Optional[Any] = False _SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name def _lowerCAmelCase ( UpperCAmelCase : Namespace ): '''simple docstring''' return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class __a ( snake_case__ ): """simple docstring""" @staticmethod def _lowerCAmelCase ( lowercase_ : ArgumentParser ): UpperCamelCase__ : Optional[Any] =parser.add_parser('''add-new-model''' ) add_new_model_parser.add_argument('''--testing''' , action='''store_true''' , help='''If in testing mode.''' ) add_new_model_parser.add_argument('''--testing_file''' , type=lowercase_ , help='''Configuration file on which to run.''' ) add_new_model_parser.add_argument( '''--path''' , type=lowercase_ , help='''Path to cookiecutter. Should only be used for testing purposes.''' ) add_new_model_parser.set_defaults(func=lowercase_ ) def __init__( self : List[Any] , lowercase_ : bool , lowercase_ : str , lowercase_ : Any=None , *lowercase_ : str ): UpperCamelCase__ : List[Any] =testing UpperCamelCase__ : Tuple =testing_file UpperCamelCase__ : Tuple =path def _lowerCAmelCase ( self : Any ): warnings.warn( '''The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ''' '''It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ''' '''checks, you should use `transformers-cli add-new-model-like` instead.''' ) if not _has_cookiecutter: raise ImportError( '''Model creation dependencies are required to use the `add_new_model` command. Install them by running ''' '''the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n''' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory UpperCamelCase__ : Tuple =[directory for directory in os.listdir() if '''cookiecutter-template-''' == directory[:22]] if len(lowercase_ ) > 0: raise ValueError( '''Several directories starting with `cookiecutter-template-` in current working directory. ''' '''Please clean your directory by removing all folders starting with `cookiecutter-template-` or ''' '''change your working directory.''' ) UpperCamelCase__ : Any =( Path(lowercase_ ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) UpperCamelCase__ : Optional[Any] =path_to_transformer_root / '''templates''' / '''adding_a_new_model''' # Execute cookiecutter if not self._testing: cookiecutter(str(lowercase_ ) ) else: with open(self._testing_file , '''r''' ) as configuration_file: UpperCamelCase__ : int =json.load(lowercase_ ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=lowercase_ , extra_context=lowercase_ , ) UpperCamelCase__ : str =[directory for directory in os.listdir() if '''cookiecutter-template-''' in directory[:22]][0] # Retrieve configuration with open(directory + '''/configuration.json''' , '''r''' ) as configuration_file: UpperCamelCase__ : int =json.load(lowercase_ ) UpperCamelCase__ : int =configuration['''lowercase_modelname'''] UpperCamelCase__ : int =configuration['''generate_tensorflow_pytorch_and_flax'''] os.remove(f'''{directory}/configuration.json''' ) UpperCamelCase__ : Dict ='''PyTorch''' in generate_tensorflow_pytorch_and_flax UpperCamelCase__ : Dict ='''TensorFlow''' in generate_tensorflow_pytorch_and_flax UpperCamelCase__ : Dict ='''Flax''' in generate_tensorflow_pytorch_and_flax UpperCamelCase__ : Union[str, Any] =f'''{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}''' os.makedirs(lowercase_ , exist_ok=lowercase_ ) os.makedirs(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}''' , exist_ok=lowercase_ ) # Tests require submodules as they have parent imports with open(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py''' , '''w''' ): pass shutil.move( f'''{directory}/__init__.py''' , f'''{model_dir}/__init__.py''' , ) shutil.move( f'''{directory}/configuration_{lowercase_model_name}.py''' , f'''{model_dir}/configuration_{lowercase_model_name}.py''' , ) def remove_copy_lines(lowercase_ : Optional[int] ): with open(lowercase_ , '''r''' ) as f: UpperCamelCase__ : Union[str, Any] =f.readlines() with open(lowercase_ , '''w''' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(lowercase_ ) if output_pytorch: if not self._testing: remove_copy_lines(f'''{directory}/modeling_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_{lowercase_model_name}.py''' ) if output_tensorflow: if not self._testing: remove_copy_lines(f'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_tf_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_tf_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' ) if output_flax: if not self._testing: remove_copy_lines(f'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_flax_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_flax_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/{lowercase_model_name}.md''' , f'''{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md''' , ) shutil.move( f'''{directory}/tokenization_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/tokenization_fast_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}_fast.py''' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(lowercase_ : str , lowercase_ : str , lowercase_ : List[str] ): # Create temp file UpperCamelCase__ , UpperCamelCase__ : Any =mkstemp() UpperCamelCase__ : Dict =False with fdopen(lowercase_ , '''w''' ) as new_file: with open(lowercase_ ) as old_file: for line in old_file: new_file.write(lowercase_ ) if line_to_copy_below in line: UpperCamelCase__ : Union[str, Any] =True for line_to_copy in lines_to_copy: new_file.write(lowercase_ ) if not line_found: raise ValueError(f'''Line {line_to_copy_below} was not found in file.''' ) # Copy the file permissions from the old file to the new file copymode(lowercase_ , lowercase_ ) # Remove original file remove(lowercase_ ) # Move new file move(lowercase_ , lowercase_ ) def skip_units(lowercase_ : Optional[Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(lowercase_ : Any ): with open(lowercase_ ) as datafile: UpperCamelCase__ : Optional[Any] =[] UpperCamelCase__ : int =False UpperCamelCase__ : Union[str, Any] =False for line in datafile: if "# To replace in: " in line and "##" not in line: UpperCamelCase__ : Optional[int] =line.split('''"''' )[1] UpperCamelCase__ : Union[str, Any] =skip_units(lowercase_ ) elif "# Below: " in line and "##" not in line: UpperCamelCase__ : List[Any] =line.split('''"''' )[1] UpperCamelCase__ : Optional[Any] =skip_units(lowercase_ ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(lowercase_ , lowercase_ , lowercase_ ) UpperCamelCase__ : Dict =[] elif "# Replace with" in line and "##" not in line: UpperCamelCase__ : int =[] elif "##" not in line: lines_to_copy.append(lowercase_ ) remove(lowercase_ ) replace_in_files(f'''{directory}/to_replace_{lowercase_model_name}.py''' ) os.rmdir(lowercase_ )
157
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
157
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device __A =False class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): pass @nightly @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_( self ) -> str: lowerCamelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion" ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = "A painting of a squirrel eating a burger " lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=lowercase , generator=lowercase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowercase ) lowerCamelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained(lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = generator.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=lowercase , generator=lowercase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = "A painting of a squirrel eating a burger " lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=lowercase , generator=lowercase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images lowerCamelCase_ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase_ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
19
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path", type=__A, default="data/dump.txt", help="The path to the data." ) parser.add_argument("--tokenizer_type", type=__A, default="bert", choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name", type=__A, default="bert-base-uncased", help="The tokenizer to use." ) parser.add_argument("--dump_file", type=__A, default="data/dump", help="The dump file prefix." ) UpperCAmelCase__ = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": UpperCAmelCase__ = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `[CLS]` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCAmelCase__ = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `<s>` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": UpperCAmelCase__ = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` UpperCAmelCase__ = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path, "r", encoding="utf8" ) as fp: UpperCAmelCase__ = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__A )} examples to process.""" ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 10_000 UpperCAmelCase__ = time.time() for text in data: UpperCAmelCase__ = f"""{bos} {text.strip()} {sep}""" UpperCAmelCase__ = tokenizer.encode(__A, add_special_tokens=__A ) rslt.append(__A ) iter += 1 if iter % interval == 0: UpperCAmelCase__ = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) UpperCAmelCase__ = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__A )} examples processed.""" ) UpperCAmelCase__ = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" UpperCAmelCase__ = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCAmelCase__ = [np.uintaa(__A ) for d in rslt] else: UpperCAmelCase__ = [np.intaa(__A ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__A, "wb" ) as handle: pickle.dump(rslt_, __A, protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
65
0
from abc import ABC, abstractmethod from typing import List, Optional class __lowerCAmelCase ( snake_case__): def __init__( self: str ): self.test() def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :Tuple = 0 lowercase :str = False while not completed: if counter == 1: self.reset() lowercase :List[str] = self.advance() if not self.does_advance(_A ): raise Exception( "Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true." ) lowercase , lowercase , lowercase :Optional[Any] = self.update(_A ) counter += 1 if counter > 1_00_00: raise Exception("update() does not fulfill the constraint." ) if self.remaining() != 0: raise Exception("Custom Constraint is not defined correctly." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: Optional[int] ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: Tuple ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: Any ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: Any ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: Any ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def SCREAMING_SNAKE_CASE ( self: List[str] , _lowerCAmelCase: Optional[int]=False ): raise NotImplementedError( F"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) class __lowerCAmelCase ( snake_case__): def __init__( self: int , _lowerCAmelCase: Optional[int] ): super(_A , self ).__init__() if not isinstance(_A , _A ) or len(_A ) == 0: raise ValueError(F"`token_ids` has to be a non-empty list, but is {token_ids}." ) if any((not isinstance(_A , _A ) or token_id < 0) for token_id in token_ids ): raise ValueError(F"Each list in `token_ids` has to be a list of positive integers, but is {token_ids}." ) lowercase :Tuple = token_ids lowercase :Tuple = len(self.token_ids ) lowercase :Optional[int] = -1 # the index of the currently fulfilled step lowercase :List[str] = False def SCREAMING_SNAKE_CASE ( self: Optional[int] ): if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: Dict ): if not isinstance(_A , _A ): raise ValueError(F"`token_id` has to be an `int`, but is {token_id} of type {type(_A )}" ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def SCREAMING_SNAKE_CASE ( self: Optional[Any] , _lowerCAmelCase: Optional[Any] ): if not isinstance(_A , _A ): raise ValueError(F"`token_id` has to be an `int`, but is {token_id} of type {type(_A )}" ) lowercase :List[Any] = False lowercase :Any = False lowercase :Any = False if self.does_advance(_A ): self.fulfilled_idx += 1 lowercase :Dict = True if self.fulfilled_idx == (self.seqlen - 1): lowercase :Optional[Any] = True lowercase :str = completed else: # failed to make progress. lowercase :Optional[Any] = True self.reset() return stepped, completed, reset def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): lowercase :Tuple = False lowercase :Any = 0 def SCREAMING_SNAKE_CASE ( self: Optional[int] ): return self.seqlen - (self.fulfilled_idx + 1) def SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _lowerCAmelCase: List[str]=False ): lowercase :Union[str, Any] = PhrasalConstraint(self.token_ids ) if stateful: lowercase :str = self.seqlen lowercase :int = self.fulfilled_idx lowercase :List[str] = self.completed return new_constraint class __lowerCAmelCase : def __init__( self: Union[str, Any] , _lowerCAmelCase: Any , _lowerCAmelCase: Dict=True ): lowercase :str = max([len(_A ) for one in nested_token_ids] ) lowercase :Any = {} for token_ids in nested_token_ids: lowercase :List[Any] = root for tidx, token_id in enumerate(_A ): if token_id not in level: lowercase :List[str] = {} lowercase :Tuple = level[token_id] if no_subsets and self.has_subsets(_A , _A ): raise ValueError( "Each list in `nested_token_ids` can't be a complete subset of another list, but is" F" {nested_token_ids}." ) lowercase :List[Any] = root def SCREAMING_SNAKE_CASE ( self: Optional[int] , _lowerCAmelCase: Dict ): lowercase :List[Any] = self.trie for current_token in current_seq: lowercase :Dict = start[current_token] lowercase :Tuple = list(start.keys() ) return next_tokens def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: Optional[int] ): lowercase :List[str] = self.next_tokens(_A ) return len(_A ) == 0 def SCREAMING_SNAKE_CASE ( self: Optional[int] , _lowerCAmelCase: str ): lowercase :List[Any] = list(root.values() ) if len(_A ) == 0: return 1 else: return sum([self.count_leaves(_A ) for nn in next_nodes] ) def SCREAMING_SNAKE_CASE ( self: Optional[Any] , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Tuple ): lowercase :List[str] = self.count_leaves(_A ) return len(_A ) != leaf_count class __lowerCAmelCase ( snake_case__): def __init__( self: int , _lowerCAmelCase: Tuple ): super(_A , self ).__init__() if not isinstance(_A , _A ) or len(_A ) == 0: raise ValueError(F"`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}." ) if any(not isinstance(_A , _A ) for token_ids in nested_token_ids ): raise ValueError(F"`nested_token_ids` has to be a list of lists, but is {nested_token_ids}." ) if any( any((not isinstance(_A , _A ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( F"Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}." ) lowercase :Union[str, Any] = DisjunctiveTrie(_A ) lowercase :Union[str, Any] = nested_token_ids lowercase :Any = self.trie.max_height lowercase :Tuple = [] lowercase :List[str] = False def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): lowercase :Dict = self.trie.next_tokens(self.current_seq ) if len(_A ) == 0: return None else: return token_list def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: str ): if not isinstance(_A , _A ): raise ValueError(F"`token_id` is supposed to be type `int`, but is {token_id} of type {type(_A )}" ) lowercase :Any = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def SCREAMING_SNAKE_CASE ( self: str , _lowerCAmelCase: str ): if not isinstance(_A , _A ): raise ValueError(F"`token_id` is supposed to be type `int`, but is {token_id} of type {type(_A )}" ) lowercase :Optional[int] = False lowercase :Tuple = False lowercase :List[Any] = False if self.does_advance(_A ): self.current_seq.append(_A ) lowercase :Any = True else: lowercase :Optional[Any] = True self.reset() lowercase :Union[str, Any] = self.trie.reached_leaf(self.current_seq ) lowercase :List[str] = completed return stepped, completed, reset def SCREAMING_SNAKE_CASE ( self: int ): lowercase :Optional[int] = False lowercase :Optional[int] = [] def SCREAMING_SNAKE_CASE ( self: List[Any] ): if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def SCREAMING_SNAKE_CASE ( self: Optional[int] , _lowerCAmelCase: Any=False ): lowercase :List[Any] = DisjunctiveConstraint(self.token_ids ) if stateful: lowercase :int = self.seqlen lowercase :int = self.current_seq lowercase :Union[str, Any] = self.completed return new_constraint class __lowerCAmelCase : def __init__( self: Optional[Any] , _lowerCAmelCase: Union[str, Any] ): lowercase :str = constraints # max # of steps required to fulfill a given constraint lowercase :Union[str, Any] = max([c.seqlen for c in constraints] ) lowercase :str = len(_A ) lowercase :Dict = False self.init_state() def SCREAMING_SNAKE_CASE ( self: Optional[int] ): lowercase :int = [] lowercase :Dict = None lowercase :List[str] = [constraint.copy(stateful=_A ) for constraint in self.constraints] def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): lowercase :Optional[int] = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): lowercase :Any = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" lowercase :Any = constraint.advance() if isinstance(_A , _A ): token_list.append(_A ) elif isinstance(_A , _A ): token_list.extend(_A ) else: lowercase :Tuple = self.inprogress_constraint.advance() if isinstance(_A , _A ): token_list.append(_A ) elif isinstance(_A , _A ): token_list.extend(_A ) if len(_A ) == 0: return None else: return token_list def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: Union[str, Any] ): self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint lowercase , lowercase :Any = self.add(_A ) # the entire list of constraints are fulfilled if self.completed: break def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: Optional[Any] ): if not isinstance(_A , _A ): raise ValueError(F"`token_id` should be an `int`, but is `{token_id}`." ) lowercase , lowercase :Optional[int] = False, False if self.completed: lowercase :List[str] = True lowercase :Union[str, Any] = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state lowercase , lowercase , lowercase :List[Any] = self.inprogress_constraint.update(_A ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=_A ) ) lowercase :Union[str, Any] = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) lowercase :int = None if len(self.pending_constraints ) == 0: # we're done! lowercase :Optional[int] = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(_A ): lowercase , lowercase , lowercase :List[str] = pending_constraint.update(_A ) if not stepped: raise Exception( "`constraint.update(token_id)` is not yielding incremental progress, " "even though `constraint.does_advance(token_id)` is true." ) if complete: self.complete_constraints.append(_A ) lowercase :Dict = None if not complete and stepped: lowercase :Optional[int] = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". lowercase :Tuple = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. lowercase :Any = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: Optional[int]=True ): lowercase :Tuple = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: lowercase :Tuple = [ constraint.copy(stateful=_A ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: lowercase :Optional[Any] = self.inprogress_constraint.copy(stateful=_A ) lowercase :Optional[int] = [constraint.copy() for constraint in self.pending_constraints] return new_state
364
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def UpperCAmelCase__ ( lowerCamelCase ): if is_torch_version("<", "2.0.0" ) or not hasattr(lowerCamelCase, "_dynamo" ): return False return isinstance(lowerCamelCase, torch._dynamo.eval_frame.OptimizedModule ) def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase = True ): lowercase :Optional[Any] = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) lowercase :str = is_compiled_module(lowerCamelCase ) if is_compiled: lowercase :str = model lowercase :str = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(lowerCamelCase, lowerCamelCase ): lowercase :Any = model.module if not keep_fpaa_wrapper: lowercase :List[Any] = getattr(lowerCamelCase, "forward" ) lowercase :Union[str, Any] = model.__dict__.pop("_original_forward", lowerCamelCase ) if original_forward is not None: while hasattr(lowerCamelCase, "__wrapped__" ): lowercase :Tuple = forward.__wrapped__ if forward == original_forward: break lowercase :Tuple = forward if getattr(lowerCamelCase, "_converted_to_transformer_engine", lowerCamelCase ): convert_model(lowerCamelCase, to_transformer_engine=lowerCamelCase ) if is_compiled: lowercase :List[Any] = model lowercase :Optional[int] = compiled_model return model def UpperCAmelCase__ ( ): PartialState().wait_for_everyone() def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(lowerCamelCase, lowerCamelCase ) elif PartialState().local_process_index == 0: torch.save(lowerCamelCase, lowerCamelCase ) @contextmanager def UpperCAmelCase__ ( **lowerCamelCase ): for key, value in kwargs.items(): lowercase :List[str] = str(lowerCamelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def UpperCAmelCase__ ( lowerCamelCase ): if not hasattr(lowerCamelCase, "__qualname__" ) and not hasattr(lowerCamelCase, "__name__" ): lowercase :Optional[int] = getattr(lowerCamelCase, "__class__", lowerCamelCase ) if hasattr(lowerCamelCase, "__qualname__" ): return obj.__qualname__ if hasattr(lowerCamelCase, "__name__" ): return obj.__name__ return str(lowerCamelCase ) def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): for key, value in source.items(): if isinstance(lowerCamelCase, lowerCamelCase ): lowercase :Tuple = destination.setdefault(lowerCamelCase, {} ) merge_dicts(lowerCamelCase, lowerCamelCase ) else: lowercase :Optional[Any] = value return destination def UpperCAmelCase__ ( lowerCamelCase = None ): if port is None: lowercase :Tuple = 29500 with socket.socket(socket.AF_INET, socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
158
0
"""simple docstring""" import pytest _UpperCamelCase : Tuple = '__dummy_dataset1__' _UpperCamelCase : Dict = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def _SCREAMING_SNAKE_CASE ( __snake_case : Optional[int] , __snake_case : Any , __snake_case : Dict ): '''simple docstring''' lowercase = dataset_loading_script_name lowercase = tmp_path / 'datasets' / script_name script_dir.mkdir(parents=__snake_case ) lowercase = script_dir / f'{script_name}.py' with open(__snake_case , 'w' ) as f: f.write(__snake_case ) return str(__snake_case )
220
"""simple docstring""" import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def _SCREAMING_SNAKE_CASE ( __snake_case : List[Any] ): '''simple docstring''' lowercase = [] embed.append( ( f'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight', f'stage{idx}.patch_embed.proj.weight', ) ) embed.append( ( f'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias', f'stage{idx}.patch_embed.proj.bias', ) ) embed.append( ( f'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight', f'stage{idx}.patch_embed.norm.weight', ) ) embed.append( ( f'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias', f'stage{idx}.patch_embed.norm.bias', ) ) return embed def _SCREAMING_SNAKE_CASE ( __snake_case : Union[str, Any] , __snake_case : List[str] ): '''simple docstring''' lowercase = [] attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked', f'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked', f'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked', f'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight', f'stage{idx}.blocks.{cnt}.attn.proj_q.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias', f'stage{idx}.blocks.{cnt}.attn.proj_q.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight', f'stage{idx}.blocks.{cnt}.attn.proj_k.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias', f'stage{idx}.blocks.{cnt}.attn.proj_k.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight', f'stage{idx}.blocks.{cnt}.attn.proj_v.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias', f'stage{idx}.blocks.{cnt}.attn.proj_v.bias', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight', f'stage{idx}.blocks.{cnt}.attn.proj.weight', ) ) attention_weights.append( ( f'cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias', f'stage{idx}.blocks.{cnt}.attn.proj.bias', ) ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight', f'stage{idx}.blocks.{cnt}.mlp.fc1.weight') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias', f'stage{idx}.blocks.{cnt}.mlp.fc1.bias') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight', f'stage{idx}.blocks.{cnt}.mlp.fc2.weight') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias', f'stage{idx}.blocks.{cnt}.mlp.fc2.bias') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight', f'stage{idx}.blocks.{cnt}.norm1.weight') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias', f'stage{idx}.blocks.{cnt}.norm1.bias') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight', f'stage{idx}.blocks.{cnt}.norm2.weight') ) attention_weights.append( (f'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias', f'stage{idx}.blocks.{cnt}.norm2.bias') ) return attention_weights def _SCREAMING_SNAKE_CASE ( __snake_case : List[str] ): '''simple docstring''' lowercase = [] token.append((f'cvt.encoder.stages.{idx}.cls_token', 'stage2.cls_token') ) return token def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowercase = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def _SCREAMING_SNAKE_CASE ( __snake_case : List[Any] , __snake_case : Dict , __snake_case : List[str] , __snake_case : Union[str, Any] ): '''simple docstring''' lowercase = 'imagenet-1k-id2label.json' lowercase = 10_00 lowercase = 'huggingface/label-files' lowercase = num_labels lowercase = json.load(open(cached_download(hf_hub_url(__snake_case , __snake_case , repo_type='dataset' ) ) , 'r' ) ) lowercase = {int(__snake_case ): v for k, v in idalabel.items()} lowercase = idalabel lowercase = {v: k for k, v in idalabel.items()} lowercase = lowercase = CvtConfig(num_labels=__snake_case , idalabel=__snake_case , labelaid=__snake_case ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowercase = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowercase = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowercase = [2, 2, 20] lowercase = [3, 12, 16] lowercase = [1_92, 7_68, 10_24] lowercase = CvtForImageClassification(__snake_case ) lowercase = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowercase = image_size lowercase = torch.load(__snake_case , map_location=torch.device('cpu' ) ) lowercase = OrderedDict() lowercase = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowercase = list_of_state_dict + cls_token(__snake_case ) lowercase = list_of_state_dict + embeddings(__snake_case ) for cnt in range(config.depth[idx] ): lowercase = list_of_state_dict + attention(__snake_case , __snake_case ) lowercase = list_of_state_dict + final() for gg in list_of_state_dict: print(__snake_case ) for i in range(len(__snake_case ) ): lowercase = original_weights[list_of_state_dict[i][1]] model.load_state_dict(__snake_case ) model.save_pretrained(__snake_case ) image_processor.save_pretrained(__snake_case ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--cvt_model', default='cvt-w24', type=str, help='Name of the cvt model you\'d like to convert.', ) parser.add_argument( '--image_size', default=3_8_4, type=int, help='Input Image Size', ) parser.add_argument( '--cvt_file_name', default=R'cvtmodels\CvT-w24-384x384-IN-22k.pth', type=str, help='Input Image Size', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _UpperCamelCase : Tuple = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
220
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) _snake_case = {"""configuration_beit""": ["""BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BeitConfig""", """BeitOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ["""BeitFeatureExtractor"""] _snake_case = ["""BeitImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """BEIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BeitForImageClassification""", """BeitForMaskedImageModeling""", """BeitForSemanticSegmentation""", """BeitModel""", """BeitPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """FlaxBeitForImageClassification""", """FlaxBeitForMaskedImageModeling""", """FlaxBeitModel""", """FlaxBeitPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
201
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class lowerCAmelCase ( lowercase_ ): def __init__( self :str , _lowercase :Optional[NestedDataStructureLike[PathLike]] = None , _lowercase :Optional[NamedSplit] = None , _lowercase :Optional[Features] = None , _lowercase :str = None , _lowercase :bool = False , _lowercase :bool = False , _lowercase :Optional[int] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = path_or_paths lowercase__ = split if split or isinstance(_lowercase , _lowercase ) else "train" lowercase__ = features lowercase__ = cache_dir lowercase__ = keep_in_memory lowercase__ = streaming lowercase__ = num_proc lowercase__ = kwargs @abstractmethod def UpperCAmelCase ( self :Any ): '''simple docstring''' pass class lowerCAmelCase ( lowercase_ ): def __init__( self :List[Any] , _lowercase :Optional[Features] = None , _lowercase :str = None , _lowercase :bool = False , _lowercase :bool = False , _lowercase :Optional[int] = None , **_lowercase :Optional[int] , ): '''simple docstring''' lowercase__ = features lowercase__ = cache_dir lowercase__ = keep_in_memory lowercase__ = streaming lowercase__ = num_proc lowercase__ = kwargs @abstractmethod def UpperCAmelCase ( self :int ): '''simple docstring''' pass
201
1
"""simple docstring""" from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging _lowercase : Optional[Any] = logging.get_logger(__name__) class _UpperCAmelCase ( _lowerCAmelCase ): a__ : Any = ["input_features", "attention_mask"] def __init__( self : int , _lowercase : Tuple=80 , _lowercase : Any=1_60_00 , _lowercase : Union[str, Any]=80 , _lowercase : Any=0.0 , _lowercase : Optional[int]=True , _lowercase : Optional[Any]=True , _lowercase : List[str]=True , **_lowercase : Dict , ): super().__init__(feature_size=_lowercase , sampling_rate=_lowercase , padding_value=_lowercase , **_lowercase ) __UpperCAmelCase = num_mel_bins __UpperCAmelCase = do_ceptral_normalize __UpperCAmelCase = normalize_means __UpperCAmelCase = normalize_vars __UpperCAmelCase = True def a ( self : Dict , _lowercase : np.ndarray , ): __UpperCAmelCase = waveform * (2**15) # Kaldi compliance: 16-bit signed integers __UpperCAmelCase = torch.from_numpy(_lowercase ).unsqueeze(0 ) __UpperCAmelCase = ta_kaldi.fbank(_lowercase , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def a ( _lowercase : np.ndarray , _lowercase : int , _lowercase : Optional[bool] = True , _lowercase : Optional[bool] = True , _lowercase : float = 0.0 , ): # make sure we normalize float32 arrays if normalize_means: __UpperCAmelCase = x[:input_length].mean(axis=0 ) __UpperCAmelCase = np.subtract(_lowercase , _lowercase ) if normalize_vars: __UpperCAmelCase = x[:input_length].std(axis=0 ) __UpperCAmelCase = np.divide(_lowercase , _lowercase ) if input_length < x.shape[0]: __UpperCAmelCase = padding_value # make sure array is in float32 __UpperCAmelCase = x.astype(np.floataa ) return x def a ( self : List[str] , _lowercase : List[np.ndarray] , _lowercase : Optional[np.ndarray] = None ): __UpperCAmelCase = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(_lowercase , _lowercase , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(_lowercase , _lowercase ) ] def __call__( self : Union[str, Any] , _lowercase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , _lowercase : Union[bool, str, PaddingStrategy] = False , _lowercase : Optional[int] = None , _lowercase : bool = False , _lowercase : Optional[int] = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : Optional[int] = None , _lowercase : Optional[bool] = None , **_lowercase : Tuple , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) __UpperCAmelCase = isinstance(_lowercase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) __UpperCAmelCase = is_batched_numpy or ( isinstance(_lowercase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCAmelCase = [np.asarray(_lowercase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_lowercase , np.ndarray ): __UpperCAmelCase = np.asarray(_lowercase , dtype=np.floataa ) elif isinstance(_lowercase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCAmelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCAmelCase = [raw_speech] # extract fbank features __UpperCAmelCase = [self._extract_fbank_features(_lowercase ) for waveform in raw_speech] # convert into correct format for padding __UpperCAmelCase = BatchFeature({'''input_features''': features} ) __UpperCAmelCase = self.pad( _lowercase , padding=_lowercase , max_length=_lowercase , truncation=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) # make sure list is in array format __UpperCAmelCase = padded_inputs.get('''input_features''' ) if isinstance(input_features[0] , _lowercase ): __UpperCAmelCase = [np.asarray(_lowercase , dtype=np.floataa ) for feature in input_features] __UpperCAmelCase = padded_inputs.get('''attention_mask''' ) if attention_mask is not None: __UpperCAmelCase = [np.asarray(_lowercase , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __UpperCAmelCase = ( np.array(_lowercase , dtype=np.intaa ) if self._get_padding_strategies(_lowercase , max_length=_lowercase ) is not PaddingStrategy.DO_NOT_PAD else None ) __UpperCAmelCase = self.normalize( padded_inputs['''input_features'''] , attention_mask=_lowercase ) if return_tensors is not None: __UpperCAmelCase = padded_inputs.convert_to_tensors(_lowercase ) return padded_inputs
332
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor _lowercase : Union[str, Any] = transforms.Compose( [ transforms.Resize((2_56, 2_56)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def lowercase__ ( snake_case_ :List[Any] ): if isinstance(snake_case_ , torch.Tensor ): return image elif isinstance(snake_case_ , PIL.Image.Image ): __UpperCAmelCase = [image] __UpperCAmelCase = [trans(img.convert('''RGB''' ) ) for img in image] __UpperCAmelCase = torch.stack(snake_case_ ) return image class _UpperCAmelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowercase : str , _lowercase : str ): super().__init__() # make sure scheduler can always be converted to DDIM __UpperCAmelCase = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=_lowercase , scheduler=_lowercase ) def a ( self : int , _lowercase : List[str] ): if strength < 0 or strength > 1: raise ValueError(F'''The value of strength should in [0.0, 1.0] but is {strength}''' ) def a ( self : List[Any] , _lowercase : List[Any] , _lowercase : Optional[Any] , _lowercase : int ): # get the original timestep using init_timestep __UpperCAmelCase = min(int(num_inference_steps * strength ) , _lowercase ) __UpperCAmelCase = max(num_inference_steps - init_timestep , 0 ) __UpperCAmelCase = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def a ( self : Optional[Any] , _lowercase : Union[str, Any] , _lowercase : Union[str, Any] , _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Tuple , _lowercase : Optional[int]=None ): if not isinstance(_lowercase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_lowercase )}''' ) __UpperCAmelCase = image.to(device=_lowercase , dtype=_lowercase ) if isinstance(_lowercase , _lowercase ) and len(_lowercase ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(_lowercase )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) __UpperCAmelCase = init_latents.shape __UpperCAmelCase = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) # get latents print('''add noise to latents at timestep''' , _lowercase ) __UpperCAmelCase = self.scheduler.add_noise(_lowercase , _lowercase , _lowercase ) __UpperCAmelCase = init_latents return latents @torch.no_grad() def __call__( self : Any , _lowercase : Union[torch.FloatTensor, PIL.Image.Image] = None , _lowercase : float = 0.8 , _lowercase : int = 1 , _lowercase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase : float = 0.0 , _lowercase : int = 50 , _lowercase : Optional[bool] = None , _lowercase : Optional[str] = "pil" , _lowercase : bool = True , ): self.check_inputs(_lowercase ) # 2. Preprocess image __UpperCAmelCase = preprocess(_lowercase ) # 3. set timesteps self.scheduler.set_timesteps(_lowercase , device=self.device ) __UpperCAmelCase , __UpperCAmelCase = self.get_timesteps(_lowercase , _lowercase , self.device ) __UpperCAmelCase = timesteps[:1].repeat(_lowercase ) # 4. Prepare latent variables __UpperCAmelCase = self.prepare_latents(_lowercase , _lowercase , _lowercase , self.unet.dtype , self.device , _lowercase ) __UpperCAmelCase = latents # 5. Denoising loop for t in self.progress_bar(_lowercase ): # 1. predict noise model_output __UpperCAmelCase = self.unet(_lowercase , _lowercase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __UpperCAmelCase = self.scheduler.step( _lowercase , _lowercase , _lowercase , eta=_lowercase , use_clipped_model_output=_lowercase , generator=_lowercase , ).prev_sample __UpperCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) __UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __UpperCAmelCase = self.numpy_to_pil(_lowercase ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=_lowercase )
332
1
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _snake_case ( lowercase__): UpperCamelCase__ : str =["""image_processor""", """tokenizer"""] UpperCamelCase__ : Union[str, Any] ="""OwlViTImageProcessor""" UpperCamelCase__ : Union[str, Any] =("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : str, __lowercase : int=None, __lowercase : str=None, **__lowercase : str ): lowercase__ = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead.", __lowercase, ) lowercase__ = kwargs.pop("feature_extractor" ) lowercase__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__lowercase, __lowercase ) def __call__( self : Any, __lowercase : Optional[Any]=None, __lowercase : List[Any]=None, __lowercase : Optional[int]=None, __lowercase : Dict="max_length", __lowercase : Any="np", **__lowercase : Optional[Any] ): if text is None and query_images is None and images is None: raise ValueError( "You have to specify at least one text or query image or image. All three cannot be none." ) if text is not None: if isinstance(__lowercase, __lowercase ) or (isinstance(__lowercase, __lowercase ) and not isinstance(text[0], __lowercase )): lowercase__ = [self.tokenizer(__lowercase, padding=__lowercase, return_tensors=__lowercase, **__lowercase )] elif isinstance(__lowercase, __lowercase ) and isinstance(text[0], __lowercase ): lowercase__ = [] # Maximum number of queries across batch lowercase__ = max([len(__lowercase ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(__lowercase ) != max_num_queries: lowercase__ = t + [" "] * (max_num_queries - len(__lowercase )) lowercase__ = self.tokenizer(__lowercase, padding=__lowercase, return_tensors=__lowercase, **__lowercase ) encodings.append(__lowercase ) else: raise TypeError("Input text should be a string, a list of strings or a nested list of strings" ) if return_tensors == "np": lowercase__ = np.concatenate([encoding["input_ids"] for encoding in encodings], axis=0 ) lowercase__ = np.concatenate([encoding["attention_mask"] for encoding in encodings], axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp lowercase__ = jnp.concatenate([encoding["input_ids"] for encoding in encodings], axis=0 ) lowercase__ = jnp.concatenate([encoding["attention_mask"] for encoding in encodings], axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch lowercase__ = torch.cat([encoding["input_ids"] for encoding in encodings], dim=0 ) lowercase__ = torch.cat([encoding["attention_mask"] for encoding in encodings], dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf lowercase__ = tf.stack([encoding["input_ids"] for encoding in encodings], axis=0 ) lowercase__ = tf.stack([encoding["attention_mask"] for encoding in encodings], axis=0 ) else: raise ValueError("Target return tensor type could not be returned" ) lowercase__ = BatchEncoding() lowercase__ = input_ids lowercase__ = attention_mask if query_images is not None: lowercase__ = BatchEncoding() lowercase__ = self.image_processor( __lowercase, return_tensors=__lowercase, **__lowercase ).pixel_values lowercase__ = query_pixel_values if images is not None: lowercase__ = self.image_processor(__lowercase, return_tensors=__lowercase, **__lowercase ) if text is not None and images is not None: lowercase__ = image_features.pixel_values return encoding elif query_images is not None and images is not None: lowercase__ = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__lowercase ), tensor_type=__lowercase ) def A__ ( self : Any, *__lowercase : List[str], **__lowercase : Optional[int] ): return self.image_processor.post_process(*__lowercase, **__lowercase ) def A__ ( self : List[str], *__lowercase : List[Any], **__lowercase : List[str] ): return self.image_processor.post_process_object_detection(*__lowercase, **__lowercase ) def A__ ( self : List[Any], *__lowercase : List[Any], **__lowercase : Union[str, Any] ): return self.image_processor.post_process_image_guided_detection(*__lowercase, **__lowercase ) def A__ ( self : Dict, *__lowercase : Union[str, Any], **__lowercase : str ): return self.tokenizer.batch_decode(*__lowercase, **__lowercase ) def A__ ( self : List[Any], *__lowercase : Optional[Any], **__lowercase : List[Any] ): return self.tokenizer.decode(*__lowercase, **__lowercase ) @property def A__ ( self : List[Any] ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.", __lowercase, ) return self.image_processor_class @property def A__ ( self : List[str] ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.", __lowercase, ) return self.image_processor
224
from __future__ import annotations def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return len(set(SCREAMING_SNAKE_CASE_ ) ) == len(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod()
224
1
import math def _SCREAMING_SNAKE_CASE ( lowercase : int ): '''simple docstring''' if not isinstance(lowercase , lowercase ): lowerCamelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if number < 1: lowerCamelCase_ = f"""Input value of [number={number}] must be > 0""" raise ValueError(lowercase ) elif number == 1: return 3 elif number == 2: return 5 else: lowerCamelCase_ = int(math.log(number // 3 , 2 ) ) + 2 lowerCamelCase_ = [3, 5] lowerCamelCase_ = 2 lowerCamelCase_ = 3 for block in range(1 , lowercase ): for _ in range(lowercase ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): lowerCamelCase : List[Any] = 0 try: lowerCamelCase : Optional[Any] = proth(number) except ValueError: print(F"""ValueError: there is no {number}th Proth number""") continue print(F"""The {number}th Proth number: {value}""")
204
from collections.abc import Callable import numpy as np def _SCREAMING_SNAKE_CASE ( lowercase : Callable , lowercase : float , lowercase : float , lowercase : float , lowercase : float ): '''simple docstring''' lowerCamelCase_ = int(np.ceil((x_end - xa) / step_size ) ) lowerCamelCase_ = np.zeros((n + 1,) ) lowerCamelCase_ = ya lowerCamelCase_ = xa for k in range(lowercase ): lowerCamelCase_ = y[k] + step_size * ode_func(lowercase , y[k] ) lowerCamelCase_ = y[k] + ( (step_size / 2) * (ode_func(lowercase , y[k] ) + ode_func(x + step_size , lowercase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
204
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __UpperCAmelCase : List[str] = logging.get_logger(__name__) __UpperCAmelCase : List[str] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase : Union[str, Any] = { "tokenizer_file": { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json", }, } __UpperCAmelCase : Any = { "gpt-neox-20b": 2048, } class UpperCAmelCase_ ( _a): '''simple docstring''' __UpperCamelCase : Tuple = VOCAB_FILES_NAMES __UpperCamelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : int = ["input_ids", "attention_mask"] def __init__( self , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE="<|endoftext|>" , __SCREAMING_SNAKE_CASE="<|endoftext|>" , __SCREAMING_SNAKE_CASE="<|endoftext|>" , __SCREAMING_SNAKE_CASE=False , **__SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) UpperCamelCase : Union[str, Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , __SCREAMING_SNAKE_CASE ) != add_prefix_space: UpperCamelCase : Optional[int] = getattr(__SCREAMING_SNAKE_CASE , pre_tok_state.pop('''type''' ) ) UpperCamelCase : Optional[int] = add_prefix_space UpperCamelCase : Optional[int] = pre_tok_class(**__SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[int] = add_prefix_space def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" UpperCamelCase : List[Any] = self._tokenizer.model.save(__SCREAMING_SNAKE_CASE , name=__SCREAMING_SNAKE_CASE ) return tuple(__SCREAMING_SNAKE_CASE ) def _lowercase ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase : Dict = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) + [self.eos_token_id] ) if len(__SCREAMING_SNAKE_CASE ) > self.model_max_length: UpperCamelCase : Optional[Any] = input_ids[-self.model_max_length :] return input_ids
352
def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] ): """simple docstring""" UpperCamelCase : list[list[float]] = [] for data in source_data: for i, el in enumerate(SCREAMING_SNAKE_CASE_ ): if len(SCREAMING_SNAKE_CASE_ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(SCREAMING_SNAKE_CASE_ ) ) return data_lists def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] , SCREAMING_SNAKE_CASE_ : list[int] ): """simple docstring""" UpperCamelCase : list[list[float]] = [] for dlist, weight in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : Union[str, Any] = min(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = max(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : list[float] = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: UpperCamelCase : Dict = F"""Invalid weight of {weight:f} provided""" raise ValueError(SCREAMING_SNAKE_CASE_ ) score_lists.append(SCREAMING_SNAKE_CASE_ ) return score_lists def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] ): """simple docstring""" UpperCamelCase : list[float] = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = final_scores[j] + ele return final_scores def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] , SCREAMING_SNAKE_CASE_ : list[int] ): """simple docstring""" UpperCamelCase : str = get_data(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = calculate_each_score(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = generate_final_scores(SCREAMING_SNAKE_CASE_ ) # append scores to source data for i, ele in enumerate(SCREAMING_SNAKE_CASE_ ): source_data[i].append(SCREAMING_SNAKE_CASE_ ) return source_data
315
0
from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=13, SCREAMING_SNAKE_CASE_=7, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=99, SCREAMING_SNAKE_CASE_=32, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=37, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=512, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=0.02, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=None, ) -> List[Any]: UpperCamelCase : Tuple = parent UpperCamelCase : Union[str, Any] = 13 UpperCamelCase : Union[str, Any] = 7 UpperCamelCase : Any = True UpperCamelCase : int = True UpperCamelCase : Union[str, Any] = True UpperCamelCase : int = True UpperCamelCase : Union[str, Any] = 99 UpperCamelCase : List[str] = 32 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Optional[Any] = 37 UpperCamelCase : Tuple = 'gelu' UpperCamelCase : Any = 0.1 UpperCamelCase : str = 0.1 UpperCamelCase : List[str] = 512 UpperCamelCase : str = 16 UpperCamelCase : List[Any] = 2 UpperCamelCase : Optional[Any] = 0.02 UpperCamelCase : Optional[int] = 3 UpperCamelCase : str = 4 UpperCamelCase : Tuple = None def snake_case_ ( self ) -> int: UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCamelCase : Dict = None if self.use_input_mask: UpperCamelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Optional[Any] = None if self.use_token_type_ids: UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size ) UpperCamelCase : Dict = None UpperCamelCase : Optional[Any] = None UpperCamelCase : Union[str, Any] = None if self.use_labels: UpperCamelCase : Any = ids_tensor([self.batch_size], self.type_sequence_label_size ) UpperCamelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) UpperCamelCase : Optional[int] = ids_tensor([self.batch_size], self.num_choices ) UpperCamelCase : Optional[Any] = RoFormerConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, return_dict=__UpperCAmelCase, ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCamelCase : Any = TFRoFormerModel(config=__UpperCAmelCase ) UpperCamelCase : List[str] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} UpperCamelCase : List[str] = [input_ids, input_mask] UpperCamelCase : Dict = model(__UpperCAmelCase ) UpperCamelCase : Dict = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCamelCase : int = True UpperCamelCase : str = TFRoFormerForCausalLM(config=__UpperCAmelCase ) UpperCamelCase : List[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCamelCase : List[Any] = model(__UpperCAmelCase )['logits'] self.parent.assertListEqual( list(prediction_scores.numpy().shape ), [self.batch_size, self.seq_length, self.vocab_size] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[Any]: UpperCamelCase : int = TFRoFormerForMaskedLM(config=__UpperCAmelCase ) UpperCamelCase : Optional[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCamelCase : int = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCamelCase : Optional[int] = self.num_labels UpperCamelCase : Optional[Any] = TFRoFormerForSequenceClassification(config=__UpperCAmelCase ) UpperCamelCase : Tuple = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCamelCase : Optional[Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCamelCase : str = self.num_choices UpperCamelCase : Optional[int] = TFRoFormerForMultipleChoice(config=__UpperCAmelCase ) UpperCamelCase : Any = tf.tile(tf.expand_dims(__UpperCAmelCase, 1 ), (1, self.num_choices, 1) ) UpperCamelCase : str = tf.tile(tf.expand_dims(__UpperCAmelCase, 1 ), (1, self.num_choices, 1) ) UpperCamelCase : Optional[int] = tf.tile(tf.expand_dims(__UpperCAmelCase, 1 ), (1, self.num_choices, 1) ) UpperCamelCase : Optional[Any] = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } UpperCamelCase : List[Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: UpperCamelCase : Dict = self.num_labels UpperCamelCase : Tuple = TFRoFormerForTokenClassification(config=__UpperCAmelCase ) UpperCamelCase : Tuple = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCamelCase : Union[str, Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[Any]: UpperCamelCase : List[str] = TFRoFormerForQuestionAnswering(config=__UpperCAmelCase ) UpperCamelCase : List[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCamelCase : Union[str, Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def snake_case_ ( self ) -> Tuple: UpperCamelCase : List[str] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[int] = config_and_inputs UpperCamelCase : Optional[Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class lowerCAmelCase_ ( snake_case_ , snake_case_ , unittest.TestCase ): UpperCAmelCase__ : Union[str, Any] = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase__ : Optional[int] = ( { "feature-extraction": TFRoFormerModel, "fill-mask": TFRoFormerForMaskedLM, "question-answering": TFRoFormerForQuestionAnswering, "text-classification": TFRoFormerForSequenceClassification, "text-generation": TFRoFormerForCausalLM, "token-classification": TFRoFormerForTokenClassification, "zero-shot": TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : str = False def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def snake_case_ ( self ) -> List[Any]: UpperCamelCase : Optional[int] = TFRoFormerModelTester(self ) UpperCamelCase : Dict = ConfigTester(self, config_class=__UpperCAmelCase, hidden_size=37 ) def snake_case_ ( self ) -> Any: self.config_tester.run_common_tests() def snake_case_ ( self ) -> List[Any]: UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def snake_case_ ( self ) -> Optional[int]: UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def snake_case_ ( self ) -> str: UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*__UpperCAmelCase ) def snake_case_ ( self ) -> int: UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def snake_case_ ( self ) -> Union[str, Any]: UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def snake_case_ ( self ) -> Optional[int]: UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def snake_case_ ( self ) -> Optional[Any]: UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def snake_case_ ( self ) -> str: UpperCamelCase : Dict = TFRoFormerModel.from_pretrained('junnyu/roformer_chinese_base' ) self.assertIsNotNone(__UpperCAmelCase ) @require_tf class lowerCAmelCase_ ( unittest.TestCase ): @slow def snake_case_ ( self ) -> Optional[int]: UpperCamelCase : Tuple = TFRoFormerForMaskedLM.from_pretrained('junnyu/roformer_chinese_base' ) UpperCamelCase : List[str] = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : Any = model(__UpperCAmelCase )[0] # TODO Replace vocab size UpperCamelCase : Any = 5_0000 UpperCamelCase : int = [1, 6, vocab_size] self.assertEqual(output.shape, __UpperCAmelCase ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. UpperCamelCase : Optional[Any] = tf.constant( [ [ [-0.12_05_33_41, -1.0_26_49_01, 0.29_22_19_46], [-1.5_13_37_83, 0.19_74_33, 0.15_19_06_07], [-5.0_13_54_03, -3.90_02_56, -0.84_03_87_64], ] ] ) tf.debugging.assert_near(output[:, :3, :3], __UpperCAmelCase, atol=1e-4 ) @require_tf class lowerCAmelCase_ ( unittest.TestCase ): UpperCAmelCase__ : Optional[Any] = 1E-4 def snake_case_ ( self ) -> Tuple: UpperCamelCase : Tuple = tf.constant([[4, 10]] ) UpperCamelCase : List[str] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6, embedding_dim=6 ) UpperCamelCase : List[Any] = emba(input_ids.shape ) UpperCamelCase : Any = tf.constant( [[0.00_00, 0.00_00, 0.00_00, 1.00_00, 1.00_00, 1.00_00], [0.84_15, 0.04_64, 0.00_22, 0.54_03, 0.99_89, 1.00_00]] ) tf.debugging.assert_near(__UpperCAmelCase, __UpperCAmelCase, atol=self.tolerance ) def snake_case_ ( self ) -> Optional[Any]: UpperCamelCase : Optional[Any] = tf.constant( [ [0.00_00, 0.00_00, 0.00_00, 0.00_00, 0.00_00], [0.84_15, 0.82_19, 0.80_20, 0.78_19, 0.76_17], [0.90_93, 0.93_64, 0.95_81, 0.97_49, 0.98_70], ] ) UpperCamelCase : int = TFRoFormerSinusoidalPositionalEmbedding(num_positions=512, embedding_dim=512 ) emba([2, 16, 512] ) UpperCamelCase : Optional[int] = emba.weight[:3, :5] tf.debugging.assert_near(__UpperCAmelCase, __UpperCAmelCase, atol=self.tolerance ) @require_tf class lowerCAmelCase_ ( unittest.TestCase ): UpperCAmelCase__ : Optional[int] = 1E-4 def snake_case_ ( self ) -> List[str]: UpperCamelCase : Any = tf.reshape(tf.range(2 * 12 * 16 * 64, dtype=tf.floataa ), shape=(2, 12, 16, 64) ) / 100 UpperCamelCase : Optional[Any] = -tf.reshape(tf.range(2 * 12 * 16 * 64, dtype=tf.floataa ), shape=(2, 12, 16, 64) ) / 100 UpperCamelCase : Any = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32, embedding_dim=64 ) UpperCamelCase : Union[str, Any] = embed_positions([2, 16, 768] )[None, None, :, :] UpperCamelCase , UpperCamelCase : int = TFRoFormerSelfAttention.apply_rotary_position_embeddings( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) UpperCamelCase : int = tf.constant( [ [0.00_00, 0.01_00, 0.02_00, 0.03_00, 0.04_00, 0.05_00, 0.06_00, 0.07_00], [-0.20_12, 0.88_97, 0.02_63, 0.94_01, 0.20_74, 0.94_63, 0.34_81, 0.93_43], [-1.70_57, 0.62_71, -1.21_45, 1.38_97, -0.63_03, 1.76_47, -0.11_73, 1.89_85], [-2.17_31, -1.63_97, -2.73_58, 0.28_54, -2.18_40, 1.71_83, -1.30_18, 2.48_71], [0.27_17, -3.61_73, -2.92_06, -2.19_88, -3.66_38, 0.38_58, -2.91_55, 2.29_80], [3.98_59, -2.15_80, -0.79_84, -4.49_04, -4.11_81, -2.02_52, -4.47_82, 1.12_53], ] ) UpperCamelCase : List[Any] = tf.constant( [ [0.00_00, -0.01_00, -0.02_00, -0.03_00, -0.04_00, -0.05_00, -0.06_00, -0.07_00], [0.20_12, -0.88_97, -0.02_63, -0.94_01, -0.20_74, -0.94_63, -0.34_81, -0.93_43], [1.70_57, -0.62_71, 1.21_45, -1.38_97, 0.63_03, -1.76_47, 0.11_73, -1.89_85], [2.17_31, 1.63_97, 2.73_58, -0.28_54, 2.18_40, -1.71_83, 1.30_18, -2.48_71], [-0.27_17, 3.61_73, 2.92_06, 2.19_88, 3.66_38, -0.38_58, 2.91_55, -2.29_80], [-3.98_59, 2.15_80, 0.79_84, 4.49_04, 4.11_81, 2.02_52, 4.47_82, -1.12_53], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8], __UpperCAmelCase, atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8], __UpperCAmelCase, atol=self.tolerance )
119
'''simple docstring''' def __lowercase ( __lowercase , __lowercase , __lowercase=False ) -> Union[str, Any]: '''simple docstring''' if isinstance(__lowercase , __lowercase ) and isinstance(__lowercase , __lowercase ): _A = len(set_a.intersection(__lowercase ) ) if alternative_union: _A = len(__lowercase ) + len(__lowercase ) else: _A = len(set_a.union(__lowercase ) ) return intersection / union if isinstance(__lowercase , (list, tuple) ) and isinstance(__lowercase , (list, tuple) ): _A = [element for element in set_a if element in set_b] if alternative_union: _A = len(__lowercase ) + len(__lowercase ) return len(__lowercase ) / union else: _A = set_a + [element for element in set_b if element not in set_a] return len(__lowercase ) / len(__lowercase ) return len(__lowercase ) / len(__lowercase ) return None if __name__ == "__main__": lowerCamelCase_ = {'''a''', '''b''', '''c''', '''d''', '''e'''} lowerCamelCase_ = {'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
79
0
"""simple docstring""" import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = (DDIMParallelScheduler,) UpperCAmelCase : List[Any] = (('''eta''', 0.0), ('''num_inference_steps''', 50)) def lowerCAmelCase_ ( self : Optional[int] , **_UpperCAmelCase : Dict ): _A = { 'num_train_timesteps': 1_000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'clip_sample': True, } config.update(**_UpperCAmelCase ) return config def lowerCAmelCase_ ( self : int , **_UpperCAmelCase : List[str] ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) _A , _A = 10, 0.0 _A = self.dummy_model() _A = self.dummy_sample_deter scheduler.set_timesteps(_UpperCAmelCase ) for t in scheduler.timesteps: _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample return sample def lowerCAmelCase_ ( self : List[Any] ): for timesteps in [100, 500, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCAmelCase ) def lowerCAmelCase_ ( self : int ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=_UpperCAmelCase ) _A = self.scheduler_classes[0] _A = self.get_scheduler_config(steps_offset=1 ) _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1] ) ) def lowerCAmelCase_ ( self : Union[str, Any] ): for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=_UpperCAmelCase , beta_end=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=_UpperCAmelCase ) def lowerCAmelCase_ ( self : int ): for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=_UpperCAmelCase ) def lowerCAmelCase_ ( self : int ): self.check_over_configs(thresholding=_UpperCAmelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=_UpperCAmelCase , prediction_type=_UpperCAmelCase , sample_max_value=_UpperCAmelCase , ) def lowerCAmelCase_ ( self : Optional[Any] ): for t in [1, 10, 49]: self.check_over_forward(time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : int ): for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500] ): self.check_over_forward(time_step=_UpperCAmelCase , num_inference_steps=_UpperCAmelCase ) def lowerCAmelCase_ ( self : str ): for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=_UpperCAmelCase , eta=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400 ) - 0.1_4771 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960 ) - 0.3_2460 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998 ) - 0.02 ) ) < 1E-5 def lowerCAmelCase_ ( self : Dict ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) _A , _A = 10, 0.0 scheduler.set_timesteps(_UpperCAmelCase ) _A = self.dummy_model() _A = self.dummy_sample_deter _A = self.dummy_sample_deter + 0.1 _A = self.dummy_sample_deter - 0.1 _A = samplea.shape[0] _A = torch.stack([samplea, samplea, samplea] , dim=0 ) _A = torch.arange(_UpperCAmelCase )[0:3, None].repeat(1 , _UpperCAmelCase ) _A = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) _A = scheduler.batch_step_no_noise(_UpperCAmelCase , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , _UpperCAmelCase ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 1147.7904 ) < 1E-2 assert abs(result_mean.item() - 0.4982 ) < 1E-3 def lowerCAmelCase_ ( self : int ): _A = self.full_loop() _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 172.0067 ) < 1E-2 assert abs(result_mean.item() - 0.22_3967 ) < 1E-3 def lowerCAmelCase_ ( self : Union[str, Any] ): _A = self.full_loop(prediction_type='v_prediction' ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 52.5302 ) < 1E-2 assert abs(result_mean.item() - 0.0684 ) < 1E-3 def lowerCAmelCase_ ( self : List[str] ): # We specify different beta, so that the first alpha is 0.99 _A = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01 ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 149.8295 ) < 1E-2 assert abs(result_mean.item() - 0.1951 ) < 1E-3 def lowerCAmelCase_ ( self : List[str] ): # We specify different beta, so that the first alpha is 0.99 _A = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01 ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 149.0784 ) < 1E-2 assert abs(result_mean.item() - 0.1941 ) < 1E-3
368
"""simple docstring""" from sklearn.metrics import fa_score import datasets a = ''' The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) ''' a = ''' Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`. - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives. - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {\'f1\': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results[\'f1\'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results[\'f1\'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro") >>> print(round(results[\'f1\'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {\'f1\': array([0.8, 0. , 0. ])} ''' a = ''' @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ ( datasets.Metric ): '''simple docstring''' def lowerCAmelCase_ ( self : Any ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('int32' ) ), 'references': datasets.Sequence(datasets.Value('int32' ) ), } if self.config_name == 'multilabel' else { 'predictions': datasets.Value('int32' ), 'references': datasets.Value('int32' ), } ) , reference_urls=['https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'] , ) def lowerCAmelCase_ ( self : int , _UpperCAmelCase : Any , _UpperCAmelCase : int , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : Dict=1 , _UpperCAmelCase : Union[str, Any]="binary" , _UpperCAmelCase : List[str]=None ): _A = fa_score( _UpperCAmelCase , _UpperCAmelCase , labels=_UpperCAmelCase , pos_label=_UpperCAmelCase , average=_UpperCAmelCase , sample_weight=_UpperCAmelCase ) return {"f1": float(_UpperCAmelCase ) if score.size == 1 else score}
271
0
import argparse import os import re _lowerCamelCase : List[str] = """src/transformers""" # Pattern that looks at the indentation in a line. _lowerCamelCase : Optional[Any] = re.compile(r"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. _lowerCamelCase : Tuple = re.compile(r"""^\s*\"([^\"]+)\":""") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _lowerCamelCase : List[str] = re.compile(r"""^\s*_import_structure\[\"([^\"]+)\"\]""") # Pattern that matches `"key",` and puts `key` in group 0. _lowerCamelCase : str = re.compile(r"""^\s*\"([^\"]+)\",\s*$""") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _lowerCamelCase : Any = re.compile(r"""\[([^\]]+)\]""") def SCREAMING_SNAKE_CASE ( lowercase_ ) -> int: """simple docstring""" A__ = _re_indent.search(lowercase_ ) return "" if search is None else search.groups()[0] def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_="" , lowercase_=None , lowercase_=None ) -> List[str]: """simple docstring""" A__ = 0 A__ = code.split('''\n''' ) if start_prompt is not None: while not lines[index].startswith(lowercase_ ): index += 1 A__ = ['''\n'''.join(lines[:index] )] else: A__ = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). A__ = [lines[index]] index += 1 while index < len(lowercase_ ) and (end_prompt is None or not lines[index].startswith(lowercase_ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(lowercase_ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ''' ''' ): current_block.append(lines[index] ) blocks.append('''\n'''.join(lowercase_ ) ) if index < len(lowercase_ ) - 1: A__ = [lines[index + 1]] index += 1 else: A__ = [] else: blocks.append('''\n'''.join(lowercase_ ) ) A__ = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(lowercase_ ) > 0: blocks.append('''\n'''.join(lowercase_ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(lowercase_ ): blocks.append('''\n'''.join(lines[index:] ) ) return blocks def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Tuple: """simple docstring""" def _inner(lowercase_ ): return key(lowercase_ ).lower().replace('''_''' , '''''' ) return _inner def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=None ) -> int: """simple docstring""" def noop(lowercase_ ): return x if key is None: A__ = noop # Constants are all uppercase, they go first. A__ = [obj for obj in objects if key(lowercase_ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. A__ = [obj for obj in objects if key(lowercase_ )[0].isupper() and not key(lowercase_ ).isupper()] # Functions begin with a lowercase, they go last. A__ = [obj for obj in objects if not key(lowercase_ )[0].isupper()] A__ = ignore_underscore(lowercase_ ) return sorted(lowercase_ , key=lowercase_ ) + sorted(lowercase_ , key=lowercase_ ) + sorted(lowercase_ , key=lowercase_ ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Tuple: """simple docstring""" def _replace(lowercase_ ): A__ = match.groups()[0] if "," not in imports: return f"""[{imports}]""" A__ = [part.strip().replace('''"''' , '''''' ) for part in imports.split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: A__ = keys[:-1] return "[" + ", ".join([f"""\"{k}\"""" for k in sort_objects(lowercase_ )] ) + "]" A__ = import_statement.split('''\n''' ) if len(lowercase_ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. A__ = 2 if lines[1].strip() == '''[''' else 1 A__ = [(i, _re_strip_line.search(lowercase_ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] A__ = sort_objects(lowercase_ , key=lambda lowercase_ : x[1] ) A__ = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(lowercase_ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: A__ = _re_bracket_content.sub(_replace , lines[1] ) else: A__ = [part.strip().replace('''"''' , '''''' ) for part in lines[1].split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: A__ = keys[:-1] A__ = get_indent(lines[1] ) + ''', '''.join([f"""\"{k}\"""" for k in sort_objects(lowercase_ )] ) return "\n".join(lowercase_ ) else: # Finally we have to deal with imports fitting on one line A__ = _re_bracket_content.sub(_replace , lowercase_ ) return import_statement def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=True ) -> Dict: """simple docstring""" with open(lowercase_ , encoding='''utf-8''' ) as f: A__ = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 A__ = split_code_in_indented_blocks( lowercase_ , start_prompt='''_import_structure = {''' , end_prompt='''if TYPE_CHECKING:''' ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(lowercase_ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. A__ = main_blocks[block_idx] A__ = block.split('''\n''' ) # Get to the start of the imports. A__ = 0 while line_idx < len(lowercase_ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: A__ = len(lowercase_ ) else: line_idx += 1 if line_idx >= len(lowercase_ ): continue # Ignore beginning and last line: they don't contain anything. A__ = '''\n'''.join(block_lines[line_idx:-1] ) A__ = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. A__ = split_code_in_indented_blocks(lowercase_ , indent_level=lowercase_ ) # We have two categories of import key: list or _import_structure[key].append/extend A__ = _re_direct_key if '''_import_structure = {''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. A__ = [(pattern.search(lowercase_ ).groups()[0] if pattern.search(lowercase_ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. A__ = [(i, key) for i, key in enumerate(lowercase_ ) if key is not None] A__ = [x[0] for x in sorted(lowercase_ , key=lambda lowercase_ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. A__ = 0 A__ = [] for i in range(len(lowercase_ ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: A__ = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(lowercase_ ) count += 1 # And we put our main block back together with its first and last line. A__ = '''\n'''.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(lowercase_ ): if check_only: return True else: print(f"""Overwriting {file}.""" ) with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write('''\n'''.join(lowercase_ ) ) def SCREAMING_SNAKE_CASE ( lowercase_=True ) -> int: """simple docstring""" A__ = [] for root, _, files in os.walk(lowercase_ ): if "__init__.py" in files: A__ = sort_imports(os.path.join(lowercase_ , '''__init__.py''' ) , check_only=lowercase_ ) if result: A__ = [os.path.join(lowercase_ , '''__init__.py''' )] if len(lowercase_ ) > 0: raise ValueError(f"""Would overwrite {len(lowercase_ )} files, run `make style`.""" ) if __name__ == "__main__": _lowerCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") _lowerCamelCase : Optional[Any] = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
14
import math import unittest def __lowercase ( _SCREAMING_SNAKE_CASE ) -> bool: '''simple docstring''' assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> Tuple: '''simple docstring''' self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) ,"""Zero doesn't have any positive factors, primes must have exactly two.""" ,) self.assertFalse( is_prime(1 ) ,"""One only has 1 positive factor, primes must have exactly two.""" ,) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
296
0
'''simple docstring''' from cva import destroyAllWindows, imread, imshow, waitKey def a__ ( lowercase : str ) -> Optional[int]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(lowercase ): for j in range(lowercase ): _UpperCamelCase = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image lowercase__ : Optional[int] = imread('image_data/lena.jpg', 1) # convert to its negative lowercase__ : Union[str, Any] = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
360
'''simple docstring''' import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor lowercase__ : Optional[Any] = logging.get_logger(__name__) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" def __init__( self : List[Any] , *lowerCAmelCase__ : List[str] , **lowerCAmelCase__ : List[str] ) -> None: '''simple docstring''' warnings.warn( '''The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use OwlViTImageProcessor instead.''' , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
287
0
'''simple docstring''' import warnings warnings.warn( "memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: " "`from accelerate import find_executable_batch_size` to avoid this warning.", FutureWarning, )
151
'''simple docstring''' import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor lowercase__ = logging.get_logger(__name__) class A_ ( _snake_case ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : Optional[Any] ) -> None: warnings.warn( 'The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PerceiverImageProcessor instead.' , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
151
1
"""simple docstring""" from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig UpperCAmelCase = logging.get_logger(__name__) # General docstring UpperCAmelCase = '''ResNetConfig''' # Base docstring UpperCAmelCase = '''microsoft/resnet-50''' UpperCAmelCase = [1, 2_048, 7, 7] # Image classification docstring UpperCAmelCase = '''microsoft/resnet-50''' UpperCAmelCase = '''tiger cat''' UpperCAmelCase = [ '''microsoft/resnet-50''', # See all resnet models at https://huggingface.co/models?filter=resnet ] class __magic_name__ ( nn.Module ): def __init__( self : List[str] , snake_case__ : int , snake_case__ : int , snake_case__ : int = 3 , snake_case__ : int = 1 , snake_case__ : str = "relu" ): '''simple docstring''' super().__init__() lowercase :List[str] = nn.Convad( snake_case__ , snake_case__ , kernel_size=snake_case__ , stride=snake_case__ , padding=kernel_size // 2 , bias=snake_case__ ) lowercase :Union[str, Any] = nn.BatchNormad(snake_case__ ) lowercase :Optional[Any] = ACTaFN[activation] if activation is not None else nn.Identity() def __snake_case ( self : Any , snake_case__ : Tensor ): '''simple docstring''' lowercase :Optional[Any] = self.convolution(snake_case__ ) lowercase :int = self.normalization(snake_case__ ) lowercase :Union[str, Any] = self.activation(snake_case__ ) return hidden_state class __magic_name__ ( nn.Module ): def __init__( self : Optional[int] , snake_case__ : ResNetConfig ): '''simple docstring''' super().__init__() lowercase :int = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) lowercase :Optional[int] = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) lowercase :Optional[Any] = config.num_channels def __snake_case ( self : Tuple , snake_case__ : Tensor ): '''simple docstring''' lowercase :List[Any] = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) lowercase :int = self.embedder(snake_case__ ) lowercase :List[Any] = self.pooler(snake_case__ ) return embedding class __magic_name__ ( nn.Module ): def __init__( self : Optional[int] , snake_case__ : int , snake_case__ : int , snake_case__ : int = 2 ): '''simple docstring''' super().__init__() lowercase :Dict = nn.Convad(snake_case__ , snake_case__ , kernel_size=1 , stride=snake_case__ , bias=snake_case__ ) lowercase :Union[str, Any] = nn.BatchNormad(snake_case__ ) def __snake_case ( self : Tuple , snake_case__ : Tensor ): '''simple docstring''' lowercase :Optional[int] = self.convolution(snake_case__ ) lowercase :List[str] = self.normalization(snake_case__ ) return hidden_state class __magic_name__ ( nn.Module ): def __init__( self : Optional[int] , snake_case__ : int , snake_case__ : int , snake_case__ : int = 1 , snake_case__ : str = "relu" ): '''simple docstring''' super().__init__() lowercase :Optional[int] = in_channels != out_channels or stride != 1 lowercase :Optional[int] = ( ResNetShortCut(snake_case__ , snake_case__ , stride=snake_case__ ) if should_apply_shortcut else nn.Identity() ) lowercase :Optional[int] = nn.Sequential( ResNetConvLayer(snake_case__ , snake_case__ , stride=snake_case__ ) , ResNetConvLayer(snake_case__ , snake_case__ , activation=snake_case__ ) , ) lowercase :Tuple = ACTaFN[activation] def __snake_case ( self : Dict , snake_case__ : List[str] ): '''simple docstring''' lowercase :List[Any] = hidden_state lowercase :Any = self.layer(snake_case__ ) lowercase :Any = self.shortcut(snake_case__ ) hidden_state += residual lowercase :Dict = self.activation(snake_case__ ) return hidden_state class __magic_name__ ( nn.Module ): def __init__( self : Tuple , snake_case__ : int , snake_case__ : int , snake_case__ : int = 1 , snake_case__ : str = "relu" , snake_case__ : int = 4 ): '''simple docstring''' super().__init__() lowercase :Optional[Any] = in_channels != out_channels or stride != 1 lowercase :str = out_channels // reduction lowercase :Optional[int] = ( ResNetShortCut(snake_case__ , snake_case__ , stride=snake_case__ ) if should_apply_shortcut else nn.Identity() ) lowercase :Tuple = nn.Sequential( ResNetConvLayer(snake_case__ , snake_case__ , kernel_size=1 ) , ResNetConvLayer(snake_case__ , snake_case__ , stride=snake_case__ ) , ResNetConvLayer(snake_case__ , snake_case__ , kernel_size=1 , activation=snake_case__ ) , ) lowercase :Tuple = ACTaFN[activation] def __snake_case ( self : Optional[int] , snake_case__ : Tuple ): '''simple docstring''' lowercase :str = hidden_state lowercase :List[str] = self.layer(snake_case__ ) lowercase :Tuple = self.shortcut(snake_case__ ) hidden_state += residual lowercase :Any = self.activation(snake_case__ ) return hidden_state class __magic_name__ ( nn.Module ): def __init__( self : Any , snake_case__ : ResNetConfig , snake_case__ : int , snake_case__ : int , snake_case__ : int = 2 , snake_case__ : int = 2 , ): '''simple docstring''' super().__init__() lowercase :Any = ResNetBottleNeckLayer if config.layer_type == '''bottleneck''' else ResNetBasicLayer lowercase :Optional[int] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , stride=snake_case__ , activation=config.hidden_act ) , *[layer(snake_case__ , snake_case__ , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def __snake_case ( self : str , snake_case__ : Tensor ): '''simple docstring''' lowercase :Dict = input for layer in self.layers: lowercase :Optional[Any] = layer(snake_case__ ) return hidden_state class __magic_name__ ( nn.Module ): def __init__( self : Optional[Any] , snake_case__ : ResNetConfig ): '''simple docstring''' super().__init__() lowercase :Any = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) lowercase :List[str] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(snake_case__ , config.depths[1:] ): self.stages.append(ResNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ ) ) def __snake_case ( self : List[Any] , snake_case__ : Tensor , snake_case__ : bool = False , snake_case__ : bool = True ): '''simple docstring''' lowercase :Any = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowercase :List[str] = hidden_states + (hidden_state,) lowercase :List[str] = stage_module(snake_case__ ) if output_hidden_states: lowercase :Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=snake_case__ , hidden_states=snake_case__ , ) class __magic_name__ ( __UpperCAmelCase ): __A : int = ResNetConfig __A : List[Any] = "resnet" __A : List[str] = "pixel_values" __A : int = True def __snake_case ( self : Dict , snake_case__ : Dict ): '''simple docstring''' if isinstance(snake_case__ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''' ) elif isinstance(snake_case__ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def __snake_case ( self : Optional[int] , snake_case__ : Optional[Any] , snake_case__ : Tuple=False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): lowercase :str = value UpperCAmelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`ResNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCAmelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare ResNet model outputting raw features without any specific head on top." , __UpperCAmelCase , ) class __magic_name__ ( __UpperCAmelCase ): def __init__( self : Tuple , snake_case__ : Union[str, Any] ): '''simple docstring''' super().__init__(snake_case__ ) lowercase :int = config lowercase :List[str] = ResNetEmbeddings(snake_case__ ) lowercase :Any = ResNetEncoder(snake_case__ ) lowercase :str = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __snake_case ( self : Tuple , snake_case__ : Tensor , snake_case__ : Optional[bool] = None , snake_case__ : Optional[bool] = None ): '''simple docstring''' lowercase :Dict = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase :List[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase :Optional[Any] = self.embedder(snake_case__ ) lowercase :str = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ ) lowercase :List[Any] = encoder_outputs[0] lowercase :Any = self.pooler(snake_case__ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( "\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , __UpperCAmelCase , ) class __magic_name__ ( __UpperCAmelCase ): def __init__( self : Union[str, Any] , snake_case__ : Any ): '''simple docstring''' super().__init__(snake_case__ ) lowercase :List[Any] = config.num_labels lowercase :Optional[int] = ResNetModel(snake_case__ ) # classification head lowercase :int = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __snake_case ( self : Any , snake_case__ : Optional[torch.FloatTensor] = None , snake_case__ : Optional[torch.LongTensor] = None , snake_case__ : Optional[bool] = None , snake_case__ : Optional[bool] = None , ): '''simple docstring''' lowercase :List[str] = return_dict if return_dict is not None else self.config.use_return_dict lowercase :str = self.resnet(snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ ) lowercase :Dict = outputs.pooler_output if return_dict else outputs[1] lowercase :Union[str, Any] = self.classifier(snake_case__ ) lowercase :str = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase :Optional[int] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase :List[str] = '''single_label_classification''' else: lowercase :List[Any] = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase :Optional[Any] = MSELoss() if self.num_labels == 1: lowercase :Dict = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase :List[str] = loss_fct(snake_case__ , snake_case__ ) elif self.config.problem_type == "single_label_classification": lowercase :int = CrossEntropyLoss() lowercase :Optional[int] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase :Optional[Any] = BCEWithLogitsLoss() lowercase :Optional[Any] = loss_fct(snake_case__ , snake_case__ ) if not return_dict: lowercase :Union[str, Any] = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states ) @add_start_docstrings( "\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n " , __UpperCAmelCase , ) class __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ): def __init__( self : int , snake_case__ : Optional[int] ): '''simple docstring''' super().__init__(snake_case__ ) super()._init_backbone(snake_case__ ) lowercase :List[str] = [config.embedding_size] + config.hidden_sizes lowercase :Dict = ResNetEmbeddings(snake_case__ ) lowercase :List[Any] = ResNetEncoder(snake_case__ ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(snake_case__ ) @replace_return_docstrings(output_type=snake_case__ , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : List[str] , snake_case__ : Tensor , snake_case__ : Optional[bool] = None , snake_case__ : Optional[bool] = None ): '''simple docstring''' lowercase :Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase :Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase :int = self.embedder(snake_case__ ) lowercase :int = self.encoder(snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ ) lowercase :Dict = outputs.hidden_states lowercase :Dict = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: lowercase :Dict = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=snake_case__ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=snake_case__ , )
359
"""simple docstring""" # DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class __magic_name__ ( __UpperCAmelCase ): __A : torch.FloatTensor __A : Optional[torch.FloatTensor] = None def lowerCamelCase (a_ :List[Any] , a_ :List[str]=0.9_99 , a_ :List[Any]="cosine" , ) -> str: if alpha_transform_type == "cosine": def alpha_bar_fn(a_ :Union[str, Any]): return math.cos((t + 0.0_08) / 1.0_08 * math.pi / 2) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(a_ :Tuple): return math.exp(t * -12.0) else: raise ValueError(F"""Unsupported alpha_tranform_type: {alpha_transform_type}""") lowercase :str = [] for i in range(a_): lowercase :Optional[Any] = i / num_diffusion_timesteps lowercase :Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(a_) / alpha_bar_fn(a_) , a_)) return torch.tensor(a_ , dtype=torch.floataa) class __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ): __A : Tuple = 1 @register_to_config def __init__( self : Union[str, Any] , snake_case__ : int = 1_0_0_0 , snake_case__ : float = 0.00_01 , snake_case__ : float = 0.02 , snake_case__ : str = "linear" , snake_case__ : Optional[Union[np.ndarray, List[float]]] = None , snake_case__ : bool = True , snake_case__ : bool = True , snake_case__ : int = 0 , snake_case__ : str = "epsilon" , snake_case__ : float = 1.0 , **snake_case__ : Union[str, Any] , ): '''simple docstring''' if kwargs.get('''set_alpha_to_one''' , snake_case__ ) is not None: lowercase :Any = ( '''The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.''' ) deprecate('''set_alpha_to_one''' , '''1.0.0''' , snake_case__ , standard_warn=snake_case__ ) lowercase :str = kwargs['''set_alpha_to_one'''] if trained_betas is not None: lowercase :Any = torch.tensor(snake_case__ , dtype=torch.floataa ) elif beta_schedule == "linear": lowercase :Optional[Any] = torch.linspace(snake_case__ , snake_case__ , snake_case__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. lowercase :Tuple = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , snake_case__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule lowercase :Dict = betas_for_alpha_bar(snake_case__ ) else: raise NotImplementedError(f"""{beta_schedule} does is not implemented for {self.__class__}""" ) lowercase :int = 1.0 - self.betas lowercase :int = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. lowercase :Tuple = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution lowercase :Any = 1.0 # setable values lowercase :Dict = None lowercase :int = torch.from_numpy(np.arange(0 , snake_case__ ).copy().astype(np.intaa ) ) def __snake_case ( self : List[Any] , snake_case__ : torch.FloatTensor , snake_case__ : Optional[int] = None ): '''simple docstring''' return sample def __snake_case ( self : Dict , snake_case__ : int , snake_case__ : Union[str, torch.device] = None ): '''simple docstring''' if num_inference_steps > self.config.num_train_timesteps: raise ValueError( f"""`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:""" f""" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle""" f""" maximal {self.config.num_train_timesteps} timesteps.""" ) lowercase :Any = num_inference_steps lowercase :Union[str, Any] = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 lowercase :str = (np.arange(0 , snake_case__ ) * step_ratio).round().copy().astype(np.intaa ) lowercase :Any = torch.from_numpy(snake_case__ ).to(snake_case__ ) self.timesteps += self.config.steps_offset def __snake_case ( self : List[Any] , snake_case__ : torch.FloatTensor , snake_case__ : int , snake_case__ : torch.FloatTensor , snake_case__ : float = 0.0 , snake_case__ : bool = False , snake_case__ : Optional[torch.FloatTensor] = None , snake_case__ : bool = True , ): '''simple docstring''' lowercase :Optional[Any] = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process lowercase :List[Any] = self.alphas_cumprod[timestep] lowercase :List[str] = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) lowercase :Any = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": lowercase :Any = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 lowercase :str = model_output elif self.config.prediction_type == "sample": lowercase :List[Any] = model_output lowercase :List[str] = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": lowercase :Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output lowercase :Tuple = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or""" ''' `v_prediction`''' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: lowercase :Dict = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowercase :List[Any] = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowercase :Optional[int] = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=snake_case__ , pred_original_sample=snake_case__ ) def __len__( self : Tuple ): '''simple docstring''' return self.config.num_train_timesteps
172
0
"""simple docstring""" import math def _UpperCAmelCase ( __lowerCamelCase : int ) -> bool: assert isinstance(__lowerCamelCase , __lowerCamelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(__lowerCamelCase ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def _UpperCAmelCase ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[int]=1 , **__lowerCamelCase : Tuple ) -> Dict: _snake_case = factor * value _snake_case = value while not is_prime(__lowerCamelCase ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **__lowerCamelCase ) return value
288
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCAmelCase__ : def __init__( self : Dict , _lowerCamelCase : int , _lowerCamelCase : Optional[int]=3 , _lowerCamelCase : List[str]=32 , _lowerCamelCase : Optional[int]=3 , _lowerCamelCase : Dict=10 , _lowerCamelCase : Tuple=[10, 20, 30, 40] , _lowerCamelCase : int=[1, 1, 2, 1] , _lowerCamelCase : int=True , _lowerCamelCase : Optional[int]=True , _lowerCamelCase : Optional[int]="relu" , _lowerCamelCase : List[Any]=3 , _lowerCamelCase : Dict=None , ): _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = embeddings_size _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = hidden_act _snake_case = num_labels _snake_case = scope _snake_case = len(_lowerCamelCase ) def lowercase ( self : Optional[int] ): _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.num_labels ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase ( self : Tuple ): return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowercase ( self : List[Any] , _lowerCamelCase : List[str] , _lowerCamelCase : str , _lowerCamelCase : List[Any] ): _snake_case = TFResNetModel(config=_lowerCamelCase ) _snake_case = model(_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowercase ( self : Dict , _lowerCamelCase : str , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Tuple ): _snake_case = self.num_labels _snake_case = TFResNetForImageClassification(_lowerCamelCase ) _snake_case = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase ( self : Tuple ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCAmelCase__ ( A_ , A_ , unittest.TestCase ): __a = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () __a = ( {"""feature-extraction""": TFResNetModel, """image-classification""": TFResNetForImageClassification} if is_tf_available() else {} ) __a = False __a = False __a = False __a = False __a = False def lowercase ( self : List[Any] ): _snake_case = TFResNetModelTester(self ) _snake_case = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def lowercase ( self : Tuple ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase ( self : List[Any] ): return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def lowercase ( self : Any ): pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def lowercase ( self : List[str] ): pass def lowercase ( self : int ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(_lowerCamelCase ) _snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def lowercase ( self : List[str] ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def lowercase ( self : Union[str, Any] ): def check_hidden_states_output(_lowerCamelCase : int , _lowerCamelCase : List[Any] , _lowerCamelCase : str ): _snake_case = model_class(_lowerCamelCase ) _snake_case = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _snake_case = layer_type _snake_case = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def lowercase ( self : Union[str, Any] ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def lowercase ( self : List[str] ): for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = TFResNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def _UpperCAmelCase ( ) -> Union[str, Any]: _snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCAmelCase__ ( unittest.TestCase ): @cached_property def lowercase ( self : Dict ): return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowercase ( self : List[Any] ): _snake_case = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=_lowerCamelCase , return_tensors='''tf''' ) # forward pass _snake_case = model(**_lowerCamelCase ) # verify the logits _snake_case = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) _snake_case = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _lowerCamelCase , atol=1e-4 ) )
288
1
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase :str = FlaxControlNetModel.from_pretrained( '''lllyasviel/sd-controlnet-canny''' , from_pt=__lowercase , dtype=jnp.bfloataa) __UpperCamelCase , __UpperCamelCase :Optional[int] = FlaxStableDiffusionControlNetPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , controlnet=__lowercase , from_pt=__lowercase , dtype=jnp.bfloataa) __UpperCamelCase :Optional[Any] = controlnet_params __UpperCamelCase :List[str] = '''bird''' __UpperCamelCase :Dict = jax.device_count() __UpperCamelCase :List[Any] = pipe.prepare_text_inputs([prompts] * num_samples) __UpperCamelCase :Union[str, Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png''') __UpperCamelCase :int = pipe.prepare_image_inputs([canny_image] * num_samples) __UpperCamelCase :int = jax.random.PRNGKey(0) __UpperCamelCase :Optional[int] = jax.random.split(__lowercase , jax.device_count()) __UpperCamelCase :Tuple = replicate(__lowercase) __UpperCamelCase :Optional[int] = shard(__lowercase) __UpperCamelCase :str = shard(__lowercase) __UpperCamelCase :Tuple = pipe( prompt_ids=__lowercase , image=__lowercase , params=__lowercase , prng_seed=__lowercase , num_inference_steps=50 , jit=__lowercase , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __UpperCamelCase :str = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:]) __UpperCamelCase :List[Any] = images[0, 253:256, 253:256, -1] __UpperCamelCase :Optional[Any] = jnp.asarray(jax.device_get(image_slice.flatten())) __UpperCamelCase :Union[str, Any] = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78]) print(f"""output_slice: {output_slice}""") assert jnp.abs(output_slice - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase , __UpperCamelCase :int = FlaxControlNetModel.from_pretrained( '''lllyasviel/sd-controlnet-openpose''' , from_pt=__lowercase , dtype=jnp.bfloataa) __UpperCamelCase , __UpperCamelCase :Dict = FlaxStableDiffusionControlNetPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , controlnet=__lowercase , from_pt=__lowercase , dtype=jnp.bfloataa) __UpperCamelCase :Optional[int] = controlnet_params __UpperCamelCase :Dict = '''Chef in the kitchen''' __UpperCamelCase :Any = jax.device_count() __UpperCamelCase :Optional[Any] = pipe.prepare_text_inputs([prompts] * num_samples) __UpperCamelCase :Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png''') __UpperCamelCase :List[Any] = pipe.prepare_image_inputs([pose_image] * num_samples) __UpperCamelCase :Any = jax.random.PRNGKey(0) __UpperCamelCase :Dict = jax.random.split(__lowercase , jax.device_count()) __UpperCamelCase :Union[str, Any] = replicate(__lowercase) __UpperCamelCase :Union[str, Any] = shard(__lowercase) __UpperCamelCase :Any = shard(__lowercase) __UpperCamelCase :List[Any] = pipe( prompt_ids=__lowercase , image=__lowercase , params=__lowercase , prng_seed=__lowercase , num_inference_steps=50 , jit=__lowercase , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __UpperCamelCase :List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:]) __UpperCamelCase :str = images[0, 253:256, 253:256, -1] __UpperCamelCase :Any = jnp.asarray(jax.device_get(image_slice.flatten())) __UpperCamelCase :str = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]]) print(f"""output_slice: {output_slice}""") assert jnp.abs(output_slice - expected_slice).max() < 1E-2
105
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __lowercase = logging.getLogger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : str = """token-classification""" def __init__( self , __lowercase) -> str: if type(__lowercase) == dict: __UpperCamelCase :List[Any] = Namespace(**__lowercase) __UpperCamelCase :Dict = import_module('''tasks''') try: __UpperCamelCase :str = getattr(__lowercase , hparams.task_type) __UpperCamelCase :TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""") __UpperCamelCase :Tuple = self.token_classification_task.get_labels(hparams.labels) __UpperCamelCase :Tuple = CrossEntropyLoss().ignore_index super().__init__(__lowercase , len(self.labels) , self.mode) def UpperCamelCase__ ( self , **__lowercase) -> List[Any]: return self.model(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Any: __UpperCamelCase :str = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type != "distilbert": __UpperCamelCase :Union[str, Any] = ( batch[2] if self.config.model_type in ['''bert''', '''xlnet'''] else None ) # XLM and RoBERTa don"t use token_type_ids __UpperCamelCase :Dict = self(**__lowercase) __UpperCamelCase :str = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :List[Any] = self.hparams for mode in ["train", "dev", "test"]: __UpperCamelCase :int = self._feature_file(__lowercase) if os.path.exists(__lowercase) and not args.overwrite_cache: logger.info('''Loading features from cached file %s''' , __lowercase) __UpperCamelCase :Any = torch.load(__lowercase) else: logger.info('''Creating features from dataset file at %s''' , args.data_dir) __UpperCamelCase :Any = self.token_classification_task.read_examples_from_file(args.data_dir , __lowercase) __UpperCamelCase :Union[str, Any] = self.token_classification_task.convert_examples_to_features( __lowercase , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ['''xlnet''']) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ['''xlnet'''] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=__lowercase , pad_on_left=bool(self.config.model_type in ['''xlnet''']) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info('''Saving features into cached file %s''' , __lowercase) torch.save(__lowercase , __lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = False) -> DataLoader: __UpperCamelCase :Tuple = self._feature_file(__lowercase) logger.info('''Loading features from cached file %s''' , __lowercase) __UpperCamelCase :str = torch.load(__lowercase) __UpperCamelCase :int = torch.tensor([f.input_ids for f in features] , dtype=torch.long) __UpperCamelCase :Optional[Any] = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) if features[0].token_type_ids is not None: __UpperCamelCase :str = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) else: __UpperCamelCase :Union[str, Any] = torch.tensor([0 for f in features] , dtype=torch.long) # HACK(we will not use this anymore soon) __UpperCamelCase :int = torch.tensor([f.label_ids for f in features] , dtype=torch.long) return DataLoader( TensorDataset(__lowercase , __lowercase , __lowercase , __lowercase) , batch_size=__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Dict: """Compute validation""" "" __UpperCamelCase :int = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type != "distilbert": __UpperCamelCase :Any = ( batch[2] if self.config.model_type in ['''bert''', '''xlnet'''] else None ) # XLM and RoBERTa don"t use token_type_ids __UpperCamelCase :Any = self(**__lowercase) __UpperCamelCase , __UpperCamelCase :Tuple = outputs[:2] __UpperCamelCase :List[str] = logits.detach().cpu().numpy() __UpperCamelCase :List[str] = inputs['''labels'''].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase__ ( self , __lowercase) -> List[str]: __UpperCamelCase :Tuple = torch.stack([x['''val_loss'''] for x in outputs]).mean() __UpperCamelCase :str = np.concatenate([x['''pred'''] for x in outputs] , axis=0) __UpperCamelCase :Any = np.argmax(__lowercase , axis=2) __UpperCamelCase :str = np.concatenate([x['''target'''] for x in outputs] , axis=0) __UpperCamelCase :List[str] = dict(enumerate(self.labels)) __UpperCamelCase :Tuple = [[] for _ in range(out_label_ids.shape[0])] __UpperCamelCase :Any = [[] for _ in range(out_label_ids.shape[0])] for i in range(out_label_ids.shape[0]): for j in range(out_label_ids.shape[1]): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]]) preds_list[i].append(label_map[preds[i][j]]) __UpperCamelCase :Any = { '''val_loss''': val_loss_mean, '''accuracy_score''': accuracy_score(__lowercase , __lowercase), '''precision''': precision_score(__lowercase , __lowercase), '''recall''': recall_score(__lowercase , __lowercase), '''f1''': fa_score(__lowercase , __lowercase), } __UpperCamelCase :Dict = dict(results.items()) __UpperCamelCase :List[str] = results return ret, preds_list, out_label_list def UpperCamelCase__ ( self , __lowercase) -> int: # when stable __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[Any] = self._eval_end(__lowercase) __UpperCamelCase :Tuple = ret['''log'''] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase__ ( self , __lowercase) -> int: # updating to test_epoch_end instead of deprecated test_end __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[int] = self._eval_end(__lowercase) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 __UpperCamelCase :Optional[Any] = ret['''log'''] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase__ ( __lowercase , __lowercase) -> Union[str, Any]: # Add NER specific options BaseTransformer.add_model_specific_args(__lowercase , __lowercase) parser.add_argument( '''--task_type''' , default='''NER''' , type=__lowercase , help='''Task type to fine tune in training (e.g. NER, POS, etc)''') parser.add_argument( '''--max_seq_length''' , default=128 , type=__lowercase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--labels''' , default='''''' , type=__lowercase , help='''Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.''' , ) parser.add_argument( '''--gpus''' , default=0 , type=__lowercase , help='''The number of GPUs allocated for this, it is by default 0 meaning none''' , ) parser.add_argument( '''--overwrite_cache''' , action='''store_true''' , help='''Overwrite the cached training and evaluation sets''') return parser if __name__ == "__main__": __lowercase = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __lowercase = NERTransformer.add_model_specific_args(parser, os.getcwd()) __lowercase = parser.parse_args() __lowercase = NERTransformer(args) __lowercase = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __lowercase = sorted(glob.glob(os.path.join(args.output_dir, '''checkpoint-epoch=*.ckpt'''), recursive=True)) __lowercase = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
105
1
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig _a = logging.get_logger(__name__) # General docstring _a = 'PoolFormerConfig' # Base docstring _a = 'sail/poolformer_s12' _a = [1, 512, 7, 7] # Image classification docstring _a = 'sail/poolformer_s12' _a = 'tabby, tabby cat' _a = [ 'sail/poolformer_s12', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def __a ( __lowerCamelCase, __lowerCamelCase = 0.0, __lowerCamelCase = False ): if drop_prob == 0.0 or not training: return input UpperCAmelCase_ : Optional[Any] = 1 - drop_prob UpperCAmelCase_ : Union[str, Any] = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets UpperCAmelCase_ : Union[str, Any] = keep_prob + torch.rand(__lowerCamelCase, dtype=input.dtype, device=input.device ) random_tensor.floor_() # binarize UpperCAmelCase_ : Optional[Any] = input.div(__lowerCamelCase ) * random_tensor return output class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ = None ): """simple docstring""" super().__init__() UpperCAmelCase_ : str = drop_prob def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return drop_path(lowercase_ , self.drop_prob , self.training ) def UpperCamelCase__ ( self ): """simple docstring""" return "p={}".format(self.drop_prob ) class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_=None ): """simple docstring""" super().__init__() UpperCAmelCase_ : Union[str, Any] = patch_size if isinstance(lowercase_ , collections.abc.Iterable ) else (patch_size, patch_size) UpperCAmelCase_ : Dict = stride if isinstance(lowercase_ , collections.abc.Iterable ) else (stride, stride) UpperCAmelCase_ : Optional[int] = padding if isinstance(lowercase_ , collections.abc.Iterable ) else (padding, padding) UpperCAmelCase_ : Optional[Any] = nn.Convad(lowercase_ , lowercase_ , kernel_size=lowercase_ , stride=lowercase_ , padding=lowercase_ ) UpperCAmelCase_ : Optional[int] = norm_layer(lowercase_ ) if norm_layer else nn.Identity() def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.projection(lowercase_ ) UpperCAmelCase_ : List[Any] = self.norm(lowercase_ ) return embeddings class A_ (nn.GroupNorm ): '''simple docstring''' def __init__( self , lowercase_ , **lowercase_ ): """simple docstring""" super().__init__(1 , lowercase_ , **lowercase_ ) class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : Optional[Any] = nn.AvgPoolad(lowercase_ , stride=1 , padding=pool_size // 2 , count_include_pad=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.pool(lowercase_ ) - hidden_states class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : Union[str, Any] = nn.Convad(lowercase_ , lowercase_ , 1 ) UpperCAmelCase_ : Any = nn.Convad(lowercase_ , lowercase_ , 1 ) UpperCAmelCase_ : Union[str, Any] = PoolFormerDropPath(lowercase_ ) if isinstance(config.hidden_act , lowercase_ ): UpperCAmelCase_ : List[str] = ACTaFN[config.hidden_act] else: UpperCAmelCase_ : str = config.hidden_act def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = self.conva(lowercase_ ) UpperCAmelCase_ : List[Any] = self.act_fn(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.drop(lowercase_ ) UpperCAmelCase_ : Dict = self.conva(lowercase_ ) UpperCAmelCase_ : Tuple = self.drop(lowercase_ ) return hidden_states class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : str = PoolFormerPooling(lowercase_ ) UpperCAmelCase_ : Dict = PoolFormerOutput(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = PoolFormerGroupNorm(lowercase_ ) UpperCAmelCase_ : Optional[int] = PoolFormerGroupNorm(lowercase_ ) # Useful for training neural nets UpperCAmelCase_ : Optional[Any] = PoolFormerDropPath(lowercase_ ) if drop_path > 0.0 else nn.Identity() UpperCAmelCase_ : str = config.use_layer_scale if config.use_layer_scale: UpperCAmelCase_ : Union[str, Any] = nn.Parameter( config.layer_scale_init_value * torch.ones((lowercase_) ) , requires_grad=lowercase_ ) UpperCAmelCase_ : Union[str, Any] = nn.Parameter( config.layer_scale_init_value * torch.ones((lowercase_) ) , requires_grad=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if self.use_layer_scale: UpperCAmelCase_ : Union[str, Any] = self.pooling(self.before_norm(lowercase_ ) ) UpperCAmelCase_ : Tuple = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection UpperCAmelCase_ : Optional[int] = hidden_states + self.drop_path(lowercase_ ) UpperCAmelCase_ : List[str] = () UpperCAmelCase_ : Union[str, Any] = self.output(self.after_norm(lowercase_ ) ) UpperCAmelCase_ : Optional[Any] = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection UpperCAmelCase_ : List[Any] = hidden_states + self.drop_path(lowercase_ ) UpperCAmelCase_ : List[Any] = (output,) + outputs return outputs else: UpperCAmelCase_ : Optional[int] = self.drop_path(self.pooling(self.before_norm(lowercase_ ) ) ) # First residual connection UpperCAmelCase_ : Optional[int] = pooling_output + hidden_states UpperCAmelCase_ : Union[str, Any] = () # Second residual connection inside the PoolFormerOutput block UpperCAmelCase_ : Optional[Any] = self.drop_path(self.output(self.after_norm(lowercase_ ) ) ) UpperCAmelCase_ : Optional[int] = hidden_states + layer_output UpperCAmelCase_ : Tuple = (output,) + outputs return outputs class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : Optional[int] = config # stochastic depth decay rule UpperCAmelCase_ : Union[str, Any] = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings UpperCAmelCase_ : List[str] = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) UpperCAmelCase_ : int = nn.ModuleList(lowercase_ ) # Transformer blocks UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : Dict = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers UpperCAmelCase_ : str = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( lowercase_ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(lowercase_ ) ) UpperCAmelCase_ : int = nn.ModuleList(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_=False , lowercase_=True ): """simple docstring""" UpperCAmelCase_ : int = () if output_hidden_states else None UpperCAmelCase_ : int = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): UpperCAmelCase_ , UpperCAmelCase_ : Tuple = layers # Get patch embeddings from hidden_states UpperCAmelCase_ : List[str] = embedding_layer(lowercase_ ) # Send the embeddings through the blocks for _, blk in enumerate(lowercase_ ): UpperCAmelCase_ : Tuple = blk(lowercase_ ) UpperCAmelCase_ : int = layer_outputs[0] if output_hidden_states: UpperCAmelCase_ : List[Any] = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=lowercase_ , hidden_states=lowercase_ ) class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = PoolFormerConfig SCREAMING_SNAKE_CASE__ : Optional[Any] = """poolformer""" SCREAMING_SNAKE_CASE__ : Optional[Any] = """pixel_values""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = True def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if isinstance(lowercase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(lowercase_ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def UpperCamelCase__ ( self , lowercase_ , lowercase_=False ): """simple docstring""" if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = value _a = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' _a = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n' @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" ,lowercase__ ,) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__(lowercase_ ) UpperCAmelCase_ : Tuple = config UpperCAmelCase_ : Dict = PoolFormerEncoder(lowercase_ ) # Initialize weights and apply final processing self.post_init() def UpperCamelCase__ ( self ): """simple docstring""" return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(lowercase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowercase_ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase__ ( self , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase_ : int = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) UpperCAmelCase_ : Optional[Any] = self.encoder( lowercase_ , output_hidden_states=lowercase_ , return_dict=lowercase_ , ) UpperCAmelCase_ : Union[str, Any] = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=lowercase_ , hidden_states=encoder_outputs.hidden_states , ) class A_ (nn.Module ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : List[str] = nn.Linear(config.hidden_size , config.hidden_size ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.dense(lowercase_ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ ,lowercase__ ,) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__(lowercase_ ) UpperCAmelCase_ : Tuple = config.num_labels UpperCAmelCase_ : List[Any] = PoolFormerModel(lowercase_ ) # Final norm UpperCAmelCase_ : Union[str, Any] = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head UpperCAmelCase_ : Optional[Any] = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowercase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase__ ( self , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase_ : Optional[int] = self.poolformer( lowercase_ , output_hidden_states=lowercase_ , return_dict=lowercase_ , ) UpperCAmelCase_ : Union[str, Any] = outputs[0] UpperCAmelCase_ : Union[str, Any] = self.classifier(self.norm(lowercase_ ).mean([-2, -1] ) ) UpperCAmelCase_ : str = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: UpperCAmelCase_ : Dict = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): UpperCAmelCase_ : Optional[Any] = "single_label_classification" else: UpperCAmelCase_ : List[str] = "multi_label_classification" if self.config.problem_type == "regression": UpperCAmelCase_ : List[Any] = MSELoss() if self.num_labels == 1: UpperCAmelCase_ : Tuple = loss_fct(logits.squeeze() , labels.squeeze() ) else: UpperCAmelCase_ : Optional[int] = loss_fct(lowercase_ , lowercase_ ) elif self.config.problem_type == "single_label_classification": UpperCAmelCase_ : Optional[Any] = CrossEntropyLoss() UpperCAmelCase_ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": UpperCAmelCase_ : Tuple = BCEWithLogitsLoss() UpperCAmelCase_ : str = loss_fct(lowercase_ , lowercase_ ) if not return_dict: UpperCAmelCase_ : Optional[Any] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowercase_ , logits=lowercase_ , hidden_states=outputs.hidden_states )
61
"""simple docstring""" import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class A_ (unittest.TestCase ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=30 , lowercase_=2 , lowercase_=3 , lowercase_=True , lowercase_=True , lowercase_=32 , lowercase_=5 , lowercase_=4 , lowercase_=37 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=10 , lowercase_=0.02 , ): """simple docstring""" UpperCAmelCase_ : Tuple = parent UpperCAmelCase_ : List[str] = batch_size UpperCAmelCase_ : Union[str, Any] = image_size UpperCAmelCase_ : List[str] = patch_size UpperCAmelCase_ : Union[str, Any] = num_channels UpperCAmelCase_ : Optional[int] = is_training UpperCAmelCase_ : Dict = use_labels UpperCAmelCase_ : Any = hidden_size UpperCAmelCase_ : Optional[Any] = num_hidden_layers UpperCAmelCase_ : Optional[Any] = num_attention_heads UpperCAmelCase_ : Dict = intermediate_size UpperCAmelCase_ : Optional[Any] = hidden_act UpperCAmelCase_ : Optional[Any] = hidden_dropout_prob UpperCAmelCase_ : Tuple = attention_probs_dropout_prob UpperCAmelCase_ : Dict = type_sequence_label_size UpperCAmelCase_ : Optional[Any] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase_ : Any = (image_size // patch_size) ** 2 UpperCAmelCase_ : List[str] = num_patches + 1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : Dict = ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) return config, pixel_values def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[str] = FlaxViTModel(config=lowercase_ ) UpperCAmelCase_ : int = model(lowercase_ ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase_ : Optional[Any] = (self.image_size, self.image_size) UpperCAmelCase_ : List[Any] = (self.patch_size, self.patch_size) UpperCAmelCase_ : str = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = self.type_sequence_label_size UpperCAmelCase_ : Tuple = FlaxViTForImageClassification(config=lowercase_ ) UpperCAmelCase_ : str = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase_ : Any = 1 UpperCAmelCase_ : Optional[int] = FlaxViTForImageClassification(lowercase_ ) UpperCAmelCase_ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : List[Any] = model(lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : Tuple = config_and_inputs UpperCAmelCase_ : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class A_ (lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxViTModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = model_class(lowercase_ ) UpperCAmelCase_ : Optional[int] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : List[str] = [*signature.parameters.keys()] UpperCAmelCase_ : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase_ : Union[str, Any] = self._prepare_for_class(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = model_class(lowercase_ ) @jax.jit def model_jitted(lowercase_ , **lowercase_ ): return model(pixel_values=lowercase_ , **lowercase_ ) with self.subTest("JIT Enabled" ): UpperCAmelCase_ : Union[str, Any] = model_jitted(**lowercase_ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): UpperCAmelCase_ : Tuple = model_jitted(**lowercase_ ).to_tuple() self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for jitted_output, output in zip(lowercase_ , lowercase_ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase_ : Union[str, Any] = model_class_name.from_pretrained("google/vit-base-patch16-224" ) UpperCAmelCase_ : List[str] = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(lowercase_ )
61
1
def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : Optional[int] = int(lowerCAmelCase__ ) if n_element < 1: UpperCAmelCase__ : Tuple = ValueError("""a should be a positive number""" ) raise my_error UpperCAmelCase__ : str = [1] UpperCAmelCase__ : List[Any] = (0, 0, 0) UpperCAmelCase__ : Optional[int] = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": __A =input('Enter the last number (nth term) of the Hamming Number Series: ') print('Formula of Hamming Number Series => 2^i * 3^j * 5^k') __A =hamming(int(n)) print('-----------------------------------------------------') print(f"""The list with nth numbers is: {hamming_numbers}""") print('-----------------------------------------------------')
366
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class _snake_case ( unittest.TestCase ): def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=4 , ): UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : str = batch_size UpperCAmelCase__ : Tuple = seq_length UpperCAmelCase__ : Any = is_training UpperCAmelCase__ : Tuple = use_attention_mask UpperCAmelCase__ : Optional[Any] = use_token_type_ids UpperCAmelCase__ : Optional[Any] = use_labels UpperCAmelCase__ : str = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Optional[int] = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Tuple = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase__ : List[Any] = max_position_embeddings UpperCAmelCase__ : Optional[int] = type_vocab_size UpperCAmelCase__ : str = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : str = num_choices def snake_case__ ( self): UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase__ : Union[str, Any] = None if self.use_attention_mask: UpperCAmelCase__ : Tuple = random_attention_mask([self.batch_size, self.seq_length]) UpperCAmelCase__ : Dict = None if self.use_token_type_ids: UpperCAmelCase__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) UpperCAmelCase__ : Optional[Any] = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self): UpperCAmelCase__ : List[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = config_and_inputs UpperCAmelCase__ : Dict = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class _snake_case ( a__ , unittest.TestCase ): lowerCAmelCase :Optional[int] = True lowerCAmelCase :Any = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def snake_case__ ( self): UpperCAmelCase__ : List[str] = FlaxRoFormerModelTester(self) @slow def snake_case__ ( self): for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""junnyu/roformer_chinese_small""" , from_pt=_lowerCamelCase) UpperCAmelCase__ : Dict = model(np.ones((1, 1))) self.assertIsNotNone(_lowerCamelCase) @require_flax class _snake_case ( unittest.TestCase ): @slow def snake_case__ ( self): UpperCAmelCase__ : Any = FlaxRoFormerForMaskedLM.from_pretrained("""junnyu/roformer_chinese_base""") UpperCAmelCase__ : int = jnp.array([[0, 1, 2, 3, 4, 5]]) UpperCAmelCase__ : Optional[int] = model(_lowerCamelCase)[0] UpperCAmelCase__ : Union[str, Any] = 5_0000 UpperCAmelCase__ : Any = (1, 6, vocab_size) self.assertEqual(output.shape , _lowerCamelCase) UpperCAmelCase__ : Union[str, Any] = jnp.array( [[[-0.1205, -1.0265, 0.2922], [-1.5134, 0.1974, 0.1519], [-5.0135, -3.9003, -0.8404]]]) self.assertTrue(jnp.allclose(output[:, :3, :3] , _lowerCamelCase , atol=1e-4))
283
0
def _UpperCamelCase ( snake_case__ = 1000 ) -> int: return sum(2 * a * ((a - 1) // 2) for a in range(3, n + 1 ) ) if __name__ == "__main__": print(solution())
157
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging _snake_case = logging.get_logger(__name__) class _snake_case ( _lowercase ): lowerCamelCase__: Tuple = ["input_features"] def __init__( self: Tuple , __lowerCamelCase: Union[str, Any]=80 , __lowerCamelCase: Optional[Any]=1_60_00 , __lowerCamelCase: Any=1_60 , __lowerCamelCase: Optional[int]=30 , __lowerCamelCase: List[str]=4_00 , __lowerCamelCase: Tuple=0.0 , __lowerCamelCase: Union[str, Any]=False , **__lowerCamelCase: Dict , ) -> Any: super().__init__( feature_size=__lowerCamelCase , sampling_rate=__lowerCamelCase , padding_value=__lowerCamelCase , return_attention_mask=__lowerCamelCase , **__lowerCamelCase , ) __UpperCAmelCase : int = n_fft __UpperCAmelCase : List[str] = hop_length __UpperCAmelCase : Optional[Any] = chunk_length __UpperCAmelCase : Union[str, Any] = chunk_length * sampling_rate __UpperCAmelCase : Any = self.n_samples // hop_length __UpperCAmelCase : Tuple = sampling_rate __UpperCAmelCase : List[Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__lowerCamelCase , min_frequency=0.0 , max_frequency=80_00.0 , sampling_rate=__lowerCamelCase , norm="slaney" , mel_scale="slaney" , ) def _lowerCamelCase ( self: List[str] , __lowerCamelCase: np.array ) -> np.ndarray: __UpperCAmelCase : List[Any] = spectrogram( __lowerCamelCase , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="log10" , ) __UpperCAmelCase : Union[str, Any] = log_spec[:, :-1] __UpperCAmelCase : List[Any] = np.maximum(__lowerCamelCase , log_spec.max() - 8.0 ) __UpperCAmelCase : str = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def _lowerCamelCase ( __lowerCamelCase: List[np.ndarray] , __lowerCamelCase: List[np.ndarray] , __lowerCamelCase: float = 0.0 ) -> List[np.ndarray]: if attention_mask is not None: __UpperCAmelCase : Tuple = np.array(__lowerCamelCase , np.intaa ) __UpperCAmelCase : Dict = [] for vector, length in zip(__lowerCamelCase , attention_mask.sum(-1 ) ): __UpperCAmelCase : Union[str, Any] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: __UpperCAmelCase : Dict = padding_value normed_input_values.append(__lowerCamelCase ) else: __UpperCAmelCase : Optional[int] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __call__( self: Dict , __lowerCamelCase: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __lowerCamelCase: bool = True , __lowerCamelCase: Optional[int] = None , __lowerCamelCase: Optional[Union[str, TensorType]] = None , __lowerCamelCase: Optional[bool] = None , __lowerCamelCase: Optional[str] = "max_length" , __lowerCamelCase: Optional[int] = None , __lowerCamelCase: Optional[int] = None , __lowerCamelCase: Optional[bool] = None , **__lowerCamelCase: Dict , ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' f''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' f''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) __UpperCAmelCase : List[Any] = isinstance(__lowerCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __UpperCAmelCase : Optional[int] = is_batched_numpy or ( isinstance(__lowerCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCAmelCase : Any = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : str = np.asarray(__lowerCamelCase , dtype=np.floataa ) elif isinstance(__lowerCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCAmelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCAmelCase : Optional[Any] = [np.asarray([raw_speech] ).T] __UpperCAmelCase : List[Any] = BatchFeature({"input_features": raw_speech} ) # convert into correct format for padding __UpperCAmelCase : List[str] = self.pad( __lowerCamelCase , padding=__lowerCamelCase , max_length=max_length if max_length else self.n_samples , truncation=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: __UpperCAmelCase : List[Any] = self.zero_mean_unit_var_norm( padded_inputs["input_features"] , attention_mask=padded_inputs["attention_mask"] , padding_value=self.padding_value , ) __UpperCAmelCase : str = np.stack(padded_inputs["input_features"] , axis=0 ) # make sure list is in array format __UpperCAmelCase : Any = padded_inputs.get("input_features" ).transpose(2 , 0 , 1 ) __UpperCAmelCase : Dict = [self._np_extract_fbank_features(__lowerCamelCase ) for waveform in input_features[0]] if isinstance(input_features[0] , __lowerCamelCase ): __UpperCAmelCase : str = [np.asarray(__lowerCamelCase , dtype=np.floataa ) for feature in input_features] else: __UpperCAmelCase : List[str] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) __UpperCAmelCase : int = padded_inputs["attention_mask"][:, :: self.hop_length] if return_tensors is not None: __UpperCAmelCase : List[str] = padded_inputs.convert_to_tensors(__lowerCamelCase ) return padded_inputs def _lowerCamelCase ( self: str ) -> Dict[str, Any]: __UpperCAmelCase : Tuple = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : Optional[Any] = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
157
1
'''simple docstring''' def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' snake_case_ = [] snake_case_ = [] snake_case_ = { '''^''': 3, '''*''': 2, '''/''': 2, '''%''': 2, '''+''': 1, '''-''': 1, } # Priority of each operator snake_case_ = len(__UpperCAmelCase ) if (len(__UpperCAmelCase ) > 7) else 7 # Print table header for output print( '''Symbol'''.center(8 ), '''Stack'''.center(__UpperCAmelCase ), '''Postfix'''.center(__UpperCAmelCase ), sep=''' | ''', ) print('''-''' * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(__UpperCAmelCase ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(__UpperCAmelCase ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(__UpperCAmelCase ) == 0: stack.append(__UpperCAmelCase ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(__UpperCAmelCase ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(__UpperCAmelCase ) # push x to stack print( x.center(8 ), (''''''.join(__UpperCAmelCase )).ljust(__UpperCAmelCase ), (''''''.join(__UpperCAmelCase )).ljust(__UpperCAmelCase ), sep=''' | ''', ) # Output in tabular format while len(__UpperCAmelCase ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( ''' '''.center(8 ), (''''''.join(__UpperCAmelCase )).ljust(__UpperCAmelCase ), (''''''.join(__UpperCAmelCase )).ljust(__UpperCAmelCase ), sep=''' | ''', ) # Output in tabular format return "".join(__UpperCAmelCase ) # return Postfix as str def __magic_name__ ( __UpperCAmelCase ) -> Dict: '''simple docstring''' snake_case_ = list(infix[::-1] ) # reverse the infix equation for i in range(len(__UpperCAmelCase ) ): if infix[i] == "(": snake_case_ = ''')''' # change "(" to ")" elif infix[i] == ")": snake_case_ = '''(''' # change ")" to "(" return (infix_2_postfix(''''''.join(__UpperCAmelCase ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": a : str = input('\nEnter an Infix Equation = ') # Input an Infix equation a : Optional[Any] = "".join(Infix.split()) # Remove spaces from the input print('\n\t', Infix, '(Infix) -> ', infix_2_prefix(Infix), '(Prefix)')
353
'''simple docstring''' import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py a : Union[str, Any] = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. a : Any = importlib.util.spec_from_file_location( 'transformers', os.path.join(PATH_TO_TRANSFORMERS, '__init__.py'), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) a : int = spec.loader.load_module() a : Dict = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` a : str = re.compile('\[(.+?)\]\((https://huggingface\.co/.+?)\)') a : str = { 'CLIPConfigMixin', 'DecisionTransformerConfigMixin', 'EncoderDecoderConfigMixin', 'RagConfigMixin', 'SpeechEncoderDecoderConfigMixin', 'VisionEncoderDecoderConfigMixin', 'VisionTextDualEncoderConfigMixin', } def __magic_name__ ( ) -> Any: '''simple docstring''' snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): snake_case_ = False # source code of `config_class` snake_case_ = inspect.getsource(__UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(__UpperCAmelCase ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` snake_case_ ,snake_case_ = checkpoint # verify the checkpoint name corresponds to the checkpoint link snake_case_ = F"https://huggingface.co/{ckpt_name}" if ckpt_link == ckpt_link_from_name: snake_case_ = True break snake_case_ = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: snake_case_ = '''\n'''.join(sorted(__UpperCAmelCase ) ) raise ValueError(F"The following configurations don't contain any valid checkpoint:\n{message}" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
72
0
from string import ascii_lowercase, ascii_uppercase def A_ ( A__ ) -> str: if not sentence: return "" a__ : Tuple = dict(zip(A__ , A__ ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
99
'''simple docstring''' from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def __a(SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' _lowerCAmelCase = [] for part_id in partition_order: _lowerCAmelCase = df.where(F'''SPARK_PARTITION_ID() = {part_id}''' ).collect() for row_idx, row in enumerate(SCREAMING_SNAKE_CASE_ ): expected_row_ids_and_row_dicts.append((F'''{part_id}_{row_idx}''', row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(100 ).repartition(1 ) _lowerCAmelCase = Spark(SCREAMING_SNAKE_CASE_ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(10 ).repartition(2 ) _lowerCAmelCase = [1, 0] _lowerCAmelCase = _generate_iterable_examples(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Reverse the partitions. _lowerCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): _lowerCAmelCase , _lowerCAmelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(10 ).repartition(1 ) _lowerCAmelCase = SparkExamplesIterable(SCREAMING_SNAKE_CASE_ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(SCREAMING_SNAKE_CASE_ ): assert row_id == F'''0_{i}''' assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch("numpy.random.Generator" ) as generator_mock: _lowerCAmelCase = lambda SCREAMING_SNAKE_CASE_ : x.reverse() _lowerCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(SCREAMING_SNAKE_CASE_ , [2, 1, 0] ) _lowerCAmelCase = SparkExamplesIterable(SCREAMING_SNAKE_CASE_ ).shuffle_data_sources(SCREAMING_SNAKE_CASE_ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(SCREAMING_SNAKE_CASE_ ): _lowerCAmelCase , _lowerCAmelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 _lowerCAmelCase = SparkExamplesIterable(SCREAMING_SNAKE_CASE_ ).shard_data_sources(worker_id=0 , num_workers=2 ) assert shard_it_a.n_shards == 2 _lowerCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(SCREAMING_SNAKE_CASE_ , [0, 2] ) for i, (row_id, row_dict) in enumerate(SCREAMING_SNAKE_CASE_ ): _lowerCAmelCase , _lowerCAmelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 _lowerCAmelCase = SparkExamplesIterable(SCREAMING_SNAKE_CASE_ ).shard_data_sources(worker_id=1 , num_workers=2 ) assert shard_it_a.n_shards == 2 _lowerCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(SCREAMING_SNAKE_CASE_ , [1, 3] ) for i, (row_id, row_dict) in enumerate(SCREAMING_SNAKE_CASE_ ): _lowerCAmelCase , _lowerCAmelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __a(): '''simple docstring''' _lowerCAmelCase = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() _lowerCAmelCase = spark.range(100 ).repartition(1 ) _lowerCAmelCase = Spark(SCREAMING_SNAKE_CASE_ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
158
0
"""simple docstring""" import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration __UpperCamelCase = [ # tf -> hf ('''/''', '''.'''), ('''layer_''', '''layers.'''), ('''kernel''', '''weight'''), ('''beta''', '''bias'''), ('''gamma''', '''weight'''), ('''pegasus''', '''model'''), ] __UpperCamelCase = [ ('''.output.dense''', '''.fc2'''), ('''intermediate.LayerNorm''', '''final_layer_norm'''), ('''intermediate.dense''', '''fc1'''), ] __UpperCamelCase = ( INIT_COMMON + [ ('''attention.self.LayerNorm''', '''self_attn_layer_norm'''), ('''attention.output.dense''', '''self_attn.out_proj'''), ('''attention.self''', '''self_attn'''), ('''attention.encdec.LayerNorm''', '''encoder_attn_layer_norm'''), ('''attention.encdec_output.dense''', '''encoder_attn.out_proj'''), ('''attention.encdec''', '''encoder_attn'''), ('''key''', '''k_proj'''), ('''value''', '''v_proj'''), ('''query''', '''q_proj'''), ('''decoder.LayerNorm''', '''decoder.layernorm_embedding'''), ] + END_COMMON ) __UpperCamelCase = ( INIT_COMMON + [ ('''embeddings.word_embeddings''', '''shared.weight'''), ('''embeddings.position_embeddings''', '''embed_positions.weight'''), ('''attention.self.LayerNorm''', '''self_attn_layer_norm'''), ('''attention.output.dense''', '''self_attn.output'''), ('''attention.self''', '''self_attn.self'''), ('''encoder.LayerNorm''', '''encoder.layernorm_embedding'''), ] + END_COMMON ) __UpperCamelCase = [ '''encdec/key/bias''', '''encdec/query/bias''', '''encdec/value/bias''', '''self/key/bias''', '''self/query/bias''', '''self/value/bias''', '''encdec_output/dense/bias''', '''attention/output/dense/bias''', ] def lowercase (SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : List[Any] ) -> Dict: for tf_name, hf_name in patterns: SCREAMING_SNAKE_CASE = k.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return k def lowercase (SCREAMING_SNAKE_CASE_ : dict , SCREAMING_SNAKE_CASE_ : dict ) -> BigBirdPegasusForConditionalGeneration: SCREAMING_SNAKE_CASE = BigBirdPegasusConfig(**SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = BigBirdPegasusForConditionalGeneration(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch_model.state_dict() SCREAMING_SNAKE_CASE = {} # separating decoder weights SCREAMING_SNAKE_CASE = {k: tf_weights[k] for k in tf_weights if k.startswith('pegasus/decoder' )} SCREAMING_SNAKE_CASE = {k: tf_weights[k] for k in tf_weights if not k.startswith('pegasus/decoder' )} for k, v in tqdm(decoder_weights.items() , 'tf -> hf conversion' ): SCREAMING_SNAKE_CASE = [k.endswith(SCREAMING_SNAKE_CASE_ ) for ending in KEYS_TO_IGNORE] if any(SCREAMING_SNAKE_CASE_ ): continue SCREAMING_SNAKE_CASE = DECODER_PATTERNS SCREAMING_SNAKE_CASE = rename_state_dict_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if new_k not in state_dict: raise ValueError(F'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): SCREAMING_SNAKE_CASE = v.T SCREAMING_SNAKE_CASE = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) assert v.shape == state_dict[new_k].shape, F'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' for k, v in tqdm(remaining_weights.items() , 'tf -> hf conversion' ): SCREAMING_SNAKE_CASE = [k.endswith(SCREAMING_SNAKE_CASE_ ) for ending in KEYS_TO_IGNORE] if any(SCREAMING_SNAKE_CASE_ ): continue SCREAMING_SNAKE_CASE = REMAINING_PATTERNS SCREAMING_SNAKE_CASE = rename_state_dict_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(F'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): SCREAMING_SNAKE_CASE = v.T SCREAMING_SNAKE_CASE = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, F'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' SCREAMING_SNAKE_CASE = mapping['model.embed_positions.weight'] SCREAMING_SNAKE_CASE = mapping.pop('model.embed_positions.weight' ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = torch_model.load_state_dict(SCREAMING_SNAKE_CASE_ , strict=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = [ k for k in missing if k not in [ 'final_logits_bias', 'model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight', ] ] assert unexpected_missing == [], F'no matches found for the following torch keys {unexpected_missing}' assert extra == [], F'no matches found for the following tf keys {extra}' return torch_model def lowercase (SCREAMING_SNAKE_CASE_ : Tuple ) -> Dict: SCREAMING_SNAKE_CASE = tf.train.list_variables(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = ['global_step'] for name, shape in tqdm(SCREAMING_SNAKE_CASE_ , desc='converting tf checkpoint to dict' ): SCREAMING_SNAKE_CASE = any(pat in name for pat in ignore_name ) if skip_key: continue SCREAMING_SNAKE_CASE = tf.train.load_variable(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = array return tf_weights def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : dict ) -> List[Any]: SCREAMING_SNAKE_CASE = get_tf_weights_as_numpy(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = convert_bigbird_pegasus(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) torch_model.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--tf_ckpt_path''', type=str, help='''passed to tf.train.list_variables''') parser.add_argument('''--save_dir''', default=None, type=str, help='''Path to the output PyTorch model.''') __UpperCamelCase = parser.parse_args() __UpperCamelCase = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
38
"""simple docstring""" from math import sqrt def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> int: SCREAMING_SNAKE_CASE = 0 for i in range(1 , int(sqrt(SCREAMING_SNAKE_CASE_ ) + 1 ) ): if n % i == 0 and i != sqrt(SCREAMING_SNAKE_CASE_ ): total += i + n // i elif i == sqrt(SCREAMING_SNAKE_CASE_ ): total += i return total - n def lowercase (SCREAMING_SNAKE_CASE_ : int = 1_00_00 ) -> int: SCREAMING_SNAKE_CASE = sum( i for i in range(1 , SCREAMING_SNAKE_CASE_ ) if sum_of_divisors(sum_of_divisors(SCREAMING_SNAKE_CASE_ ) ) == i and sum_of_divisors(SCREAMING_SNAKE_CASE_ ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
38
1
def lowerCAmelCase_ ( __UpperCAmelCase: str ) -> bool: UpperCamelCase__ : Optional[Any] = 0 for ch in input_str: UpperCamelCase__ : Tuple = ord(__UpperCAmelCase ) UpperCamelCase__ : Tuple = pow(2 , __UpperCAmelCase ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
201
import random def lowerCAmelCase_ ( __UpperCAmelCase: list , __UpperCAmelCase: Optional[int] ) -> tuple: UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ : List[Any] = [], [], [] for element in data: if element < pivot: less.append(__UpperCAmelCase ) elif element > pivot: greater.append(__UpperCAmelCase ) else: equal.append(__UpperCAmelCase ) return less, equal, greater def lowerCAmelCase_ ( __UpperCAmelCase: list , __UpperCAmelCase: int ) -> List[str]: # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(__UpperCAmelCase ) or index < 0: return None UpperCamelCase__ : List[str] = items[random.randint(0 , len(__UpperCAmelCase ) - 1 )] UpperCamelCase__ : List[Any] = 0 UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ : int = _partition(__UpperCAmelCase , __UpperCAmelCase ) UpperCamelCase__ : Union[str, Any] = len(__UpperCAmelCase ) UpperCamelCase__ : Dict = len(__UpperCAmelCase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__UpperCAmelCase , __UpperCAmelCase ) # must be in larger else: return quick_select(__UpperCAmelCase , index - (m + count) )
201
1
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __UpperCamelCase : @staticmethod def lowercase__ ( *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" pass @is_pipeline_test @require_vision @require_timm @require_torch class __UpperCamelCase ( unittest.TestCase ): lowercase : Dict =MODEL_FOR_OBJECT_DETECTION_MAPPING def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =ObjectDetectionPipeline(model=lowerCAmelCase, image_processor=lowerCAmelCase ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =object_detector('''./tests/fixtures/tests_samples/COCO/000000039769.png''', threshold=0.0 ) self.assertGreater(len(lowerCAmelCase ), 0 ) for detected_object in outputs: self.assertEqual( lowerCAmelCase, { '''score''': ANY(lowerCAmelCase ), '''label''': ANY(lowerCAmelCase ), '''box''': {'''xmin''': ANY(lowerCAmelCase ), '''ymin''': ANY(lowerCAmelCase ), '''xmax''': ANY(lowerCAmelCase ), '''ymax''': ANY(lowerCAmelCase )}, }, ) import datasets lowerCamelCase_ =datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''', '''image''', split='''test''' ) lowerCamelCase_ =[ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] lowerCamelCase_ =object_detector(lowerCAmelCase, threshold=0.0 ) self.assertEqual(len(lowerCAmelCase ), len(lowerCAmelCase ) ) for outputs in batch_outputs: self.assertGreater(len(lowerCAmelCase ), 0 ) for detected_object in outputs: self.assertEqual( lowerCAmelCase, { '''score''': ANY(lowerCAmelCase ), '''label''': ANY(lowerCAmelCase ), '''box''': {'''xmin''': ANY(lowerCAmelCase ), '''ymin''': ANY(lowerCAmelCase ), '''xmax''': ANY(lowerCAmelCase ), '''ymax''': ANY(lowerCAmelCase )}, }, ) @require_tf @unittest.skip('''Object detection not implemented in TF''' ) def lowercase__ ( self ): """simple docstring""" pass @require_torch def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''hf-internal-testing/tiny-detr-mobilenetsv3''' lowerCamelCase_ =AutoModelForObjectDetection.from_pretrained(lowerCAmelCase ) lowerCamelCase_ =AutoFeatureExtractor.from_pretrained(lowerCAmelCase ) lowerCamelCase_ =ObjectDetectionPipeline(model=lowerCAmelCase, feature_extractor=lowerCAmelCase ) lowerCamelCase_ =object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''', threshold=0.0 ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], ) lowerCamelCase_ =object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ], threshold=0.0, ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ [ {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], [ {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3_3_7_6, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], ], ) @require_torch @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''facebook/detr-resnet-50''' lowerCamelCase_ =AutoModelForObjectDetection.from_pretrained(lowerCAmelCase ) lowerCamelCase_ =AutoFeatureExtractor.from_pretrained(lowerCAmelCase ) lowerCamelCase_ =ObjectDetectionPipeline(model=lowerCAmelCase, feature_extractor=lowerCAmelCase ) lowerCamelCase_ =object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ) lowerCamelCase_ =object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ], ) @require_torch @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''facebook/detr-resnet-50''' lowerCamelCase_ =pipeline('''object-detection''', model=lowerCAmelCase ) lowerCamelCase_ =object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ) lowerCamelCase_ =object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9_9_8_2, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9_9_6_0, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9_9_5_5, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ], ) @require_torch @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =0.9_9_8_5 lowerCamelCase_ ='''facebook/detr-resnet-50''' lowerCamelCase_ =pipeline('''object-detection''', model=lowerCAmelCase ) lowerCamelCase_ =object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''', threshold=lowerCAmelCase ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ {'''score''': 0.9_9_8_8, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9_9_8_7, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ) @require_torch @require_pytesseract @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''Narsil/layoutlmv3-finetuned-funsd''' lowerCamelCase_ =0.9_9_9_3 lowerCamelCase_ =pipeline('''object-detection''', model=lowerCAmelCase, threshold=lowerCAmelCase ) lowerCamelCase_ =object_detector( '''https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png''' ) self.assertEqual( nested_simplify(lowerCAmelCase, decimals=4 ), [ {'''score''': 0.9_9_9_3, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, {'''score''': 0.9_9_9_3, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, ], )
6
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __UpperCamelCase ( lowerCamelCase__ ): lowercase : List[str] =['image_processor', 'tokenizer'] lowercase : Optional[int] ='AutoImageProcessor' lowercase : List[str] ='AutoTokenizer' def __init__( self, lowerCAmelCase=None, lowerCAmelCase=None, **lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''', lowerCAmelCase, ) lowerCamelCase_ =kwargs.pop('''feature_extractor''' ) lowerCamelCase_ =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCAmelCase, lowerCAmelCase ) lowerCamelCase_ =self.image_processor lowerCamelCase_ =False def __call__( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" if self._in_target_context_manager: return self.current_processor(*lowerCAmelCase, **lowerCAmelCase ) lowerCamelCase_ =kwargs.pop('''images''', lowerCAmelCase ) lowerCamelCase_ =kwargs.pop('''text''', lowerCAmelCase ) if len(lowerCAmelCase ) > 0: lowerCamelCase_ =args[0] lowerCamelCase_ =args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: lowerCamelCase_ =self.image_processor(lowerCAmelCase, *lowerCAmelCase, **lowerCAmelCase ) if text is not None: lowerCamelCase_ =self.tokenizer(lowerCAmelCase, **lowerCAmelCase ) if text is None: return inputs elif images is None: return encodings else: lowerCamelCase_ =encodings['''input_ids'''] return inputs def lowercase__ ( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*lowerCAmelCase, **lowerCAmelCase ) def lowercase__ ( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" return self.tokenizer.decode(*lowerCAmelCase, **lowerCAmelCase ) @contextmanager def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) lowerCamelCase_ =True lowerCamelCase_ =self.tokenizer yield lowerCamelCase_ =self.image_processor lowerCamelCase_ =False def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=False, lowerCAmelCase=None ): """simple docstring""" if added_vocab is None: lowerCamelCase_ =self.tokenizer.get_added_vocab() lowerCamelCase_ ={} while tokens: lowerCamelCase_ =re.search(R'''<s_(.*?)>''', lowerCAmelCase, re.IGNORECASE ) if start_token is None: break lowerCamelCase_ =start_token.group(1 ) lowerCamelCase_ =re.search(Rf'''</s_{key}>''', lowerCAmelCase, re.IGNORECASE ) lowerCamelCase_ =start_token.group() if end_token is None: lowerCamelCase_ =tokens.replace(lowerCAmelCase, '''''' ) else: lowerCamelCase_ =end_token.group() lowerCamelCase_ =re.escape(lowerCAmelCase ) lowerCamelCase_ =re.escape(lowerCAmelCase ) lowerCamelCase_ =re.search(f'''{start_token_escaped}(.*?){end_token_escaped}''', lowerCAmelCase, re.IGNORECASE ) if content is not None: lowerCamelCase_ =content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node lowerCamelCase_ =self.tokenajson(lowerCAmelCase, is_inner_value=lowerCAmelCase, added_vocab=lowerCAmelCase ) if value: if len(lowerCAmelCase ) == 1: lowerCamelCase_ =value[0] lowerCamelCase_ =value else: # leaf nodes lowerCamelCase_ =[] for leaf in content.split(R'''<sep/>''' ): lowerCamelCase_ =leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": lowerCamelCase_ =leaf[1:-2] # for categorical special tokens output[key].append(lowerCAmelCase ) if len(output[key] ) == 1: lowerCamelCase_ =output[key][0] lowerCamelCase_ =tokens[tokens.find(lowerCAmelCase ) + len(lowerCAmelCase ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=lowerCAmelCase, added_vocab=lowerCAmelCase ) if len(lowerCAmelCase ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''', lowerCAmelCase, ) return self.image_processor_class @property def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''', lowerCAmelCase, ) return self.image_processor
6
1
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : list[int] , lowerCAmelCase__ : list[int] ) -> tuple[float, float]: """simple docstring""" if not len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients lowerCAmelCase_ ,lowerCAmelCase_ ,lowerCAmelCase_ : Dict = equationa lowerCAmelCase_ ,lowerCAmelCase_ ,lowerCAmelCase_ : List[Any] = equationa # Calculate the determinants of the matrices lowerCAmelCase_ : List[str] = aa * ba - aa * ba lowerCAmelCase_ : str = ca * ba - ca * ba lowerCAmelCase_ : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: lowerCAmelCase_ : Optional[int] = determinant_x / determinant lowerCAmelCase_ : Union[str, Any] = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
224
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : int = 100 ) -> int: """simple docstring""" lowerCAmelCase_ : Any = (n * (n + 1) // 2) ** 2 lowerCAmelCase_ : Optional[int] = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
224
1
from __future__ import annotations def a( A : Dict , A : Optional[Any] = None , A : Tuple = None , A : Dict = False , ) -> Dict: """simple docstring""" a = cipher_alphabet or [chr(_lowercase ) for i in range(97 , 123 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) a = { "a": 0.08_497, "b": 0.01_492, "c": 0.02_202, "d": 0.04_253, "e": 0.11_162, "f": 0.02_228, "g": 0.02_015, "h": 0.06_094, "i": 0.07_546, "j": 0.00_153, "k": 0.01_292, "l": 0.04_025, "m": 0.02_406, "n": 0.06_749, "o": 0.07_507, "p": 0.01_929, "q": 0.00_095, "r": 0.07_587, "s": 0.06_327, "t": 0.09_356, "u": 0.02_758, "v": 0.00_978, "w": 0.02_560, "x": 0.00_150, "y": 0.01_994, "z": 0.00_077, } else: # Custom frequencies dictionary a = frequencies_dict if not case_sensitive: a = ciphertext.lower() # Chi squared statistic values a = {} # cycle through all of the shifts for shift in range(len(_lowercase ) ): a = "" # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet a = (alphabet_letters.index(letter.lower() ) - shift) % len( _lowercase ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter a = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: a = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message a = decrypted_with_shift.lower().count(_lowercase ) # Get the excepcted amount of times the letter should appear based # on letter frequencies a = frequencies[letter] * occurrences # Complete the chi squared statistic formula a = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message a = decrypted_with_shift.count(_lowercase ) # Get the excepcted amount of times the letter should appear based # on letter frequencies a = frequencies[letter] * occurrences # Complete the chi squared statistic formula a = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary a = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(A : List[str] ) -> tuple[float, str]: return chi_squared_statistic_values[key] a = min( _lowercase , key=_lowercase , ) # Get all the data from the most likely cipher (key, decoded message) ( ( a ) , ( a ) , ) = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
358
def a( A : int = 200 ) -> int: """simple docstring""" a = [1, 2, 5, 10, 20, 50, 100, 200] a = [0] * (pence + 1) a = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(A , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
71
0
from __future__ import annotations def __lowerCamelCase ( lowerCamelCase__ : tuple[int, int] , lowerCamelCase__ : int ): '''simple docstring''' lowerCamelCase , lowerCamelCase = position lowerCamelCase = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase = [] for position in positions: lowerCamelCase , lowerCamelCase = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_snake_case ) return permissible_positions def __lowerCamelCase ( lowerCamelCase__ : list[list[int]] ): '''simple docstring''' return not any(elem == 0 for row in board for elem in row ) def __lowerCamelCase ( lowerCamelCase__ : list[list[int]] , lowerCamelCase__ : tuple[int, int] , lowerCamelCase__ : int ): '''simple docstring''' if is_complete(_snake_case ): return True for position in get_valid_pos(_snake_case , len(_snake_case ) ): lowerCamelCase , lowerCamelCase = position if board[y][x] == 0: lowerCamelCase = curr + 1 if open_knight_tour_helper(_snake_case , _snake_case , curr + 1 ): return True lowerCamelCase = 0 return False def __lowerCamelCase ( lowerCamelCase__ : int ): '''simple docstring''' lowerCamelCase = [[0 for i in range(_snake_case )] for j in range(_snake_case )] for i in range(_snake_case ): for j in range(_snake_case ): lowerCamelCase = 1 if open_knight_tour_helper(_snake_case , (i, j) , 1 ): return board lowerCamelCase = 0 lowerCamelCase = f'Open Kight Tour cannot be performed on a board of size {n}' raise ValueError(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
252
"""simple docstring""" from __future__ import annotations def _snake_case ( _snake_case : tuple[int, int] , _snake_case : int ) -> list[tuple[int, int]]: '''simple docstring''' _A , _A = position _A = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] _A = [] for position in positions: _A , _A = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_snake_case ) return permissible_positions def _snake_case ( _snake_case : list[list[int]] ) -> bool: '''simple docstring''' return not any(elem == 0 for row in board for elem in row ) def _snake_case ( _snake_case : list[list[int]] , _snake_case : tuple[int, int] , _snake_case : int ) -> bool: '''simple docstring''' if is_complete(_snake_case ): return True for position in get_valid_pos(_snake_case , len(_snake_case ) ): _A , _A = position if board[y][x] == 0: _A = curr + 1 if open_knight_tour_helper(_snake_case , _snake_case , curr + 1 ): return True _A = 0 return False def _snake_case ( _snake_case : int ) -> list[list[int]]: '''simple docstring''' _A = [[0 for i in range(_snake_case )] for j in range(_snake_case )] for i in range(_snake_case ): for j in range(_snake_case ): _A = 1 if open_knight_tour_helper(_snake_case , (i, j) , 1 ): return board _A = 0 _A = F'''Open Kight Tour cannot be performed on a board of size {n}''' raise ValueError(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
315
0
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = { '''huggingface/time-series-transformer-tourism-monthly''': ( '''https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json''' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class A ( _snake_case ): UpperCamelCase_ : List[str] ='''time_series_transformer''' UpperCamelCase_ : int ={ '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__(self , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "student_t" , lowerCAmelCase = "nll" , lowerCAmelCase = 1 , lowerCAmelCase = [1, 2, 3, 4, 5, 6, 7] , lowerCAmelCase = "mean" , lowerCAmelCase = 0 , lowerCAmelCase = 0 , lowerCAmelCase = 0 , lowerCAmelCase = 0 , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = 3_2 , lowerCAmelCase = 3_2 , lowerCAmelCase = 2 , lowerCAmelCase = 2 , lowerCAmelCase = 2 , lowerCAmelCase = 2 , lowerCAmelCase = True , lowerCAmelCase = "gelu" , lowerCAmelCase = 6_4 , lowerCAmelCase = 0.1 , lowerCAmelCase = 0.1 , lowerCAmelCase = 0.1 , lowerCAmelCase = 0.1 , lowerCAmelCase = 0.1 , lowerCAmelCase = 1_0_0 , lowerCAmelCase = 0.02 , lowerCAmelCase=True , **lowerCAmelCase , ): __lowercase= prediction_length __lowercase= context_length or prediction_length __lowercase= distribution_output __lowercase= loss __lowercase= input_size __lowercase= num_time_features __lowercase= lags_sequence __lowercase= scaling __lowercase= num_dynamic_real_features __lowercase= num_static_real_features __lowercase= num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(UpperCamelCase__ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) __lowercase= cardinality else: __lowercase= [0] if embedding_dimension and num_static_categorical_features > 0: if len(UpperCamelCase__ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) __lowercase= embedding_dimension else: __lowercase= [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] __lowercase= num_parallel_samples # Transformer architecture configuration __lowercase= input_size * len(UpperCamelCase__ ) + self._number_of_features __lowercase= d_model __lowercase= encoder_attention_heads __lowercase= decoder_attention_heads __lowercase= encoder_ffn_dim __lowercase= decoder_ffn_dim __lowercase= encoder_layers __lowercase= decoder_layers __lowercase= dropout __lowercase= attention_dropout __lowercase= activation_dropout __lowercase= encoder_layerdrop __lowercase= decoder_layerdrop __lowercase= activation_function __lowercase= init_std __lowercase= use_cache super().__init__(is_encoder_decoder=UpperCamelCase__ , **UpperCamelCase__ ) @property def _A (self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
363
import math from datetime import datetime, timedelta def _lowerCamelCase( lowercase__ ) -> datetime: '''simple docstring''' __lowercase= year % 1_9 __lowercase= year % 4 __lowercase= year % 7 __lowercase= math.floor(year / 1_0_0 ) __lowercase= math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) __lowercase= leap_day_inhibits / 4 __lowercase= ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 __lowercase= (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 __lowercase= (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon __lowercase= ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(lowercase__ , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(lowercase__ , 4 , 1_8 ) else: return datetime(lowercase__ , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1_9_9_4, 2_0_0_0, 2_0_1_0, 2_0_2_1, 2_0_2_3): lowerCAmelCase = '''will be''' if year > datetime.now().year else '''was''' print(F'Easter in {year} {tense} {gauss_easter(year)}')
304
0
import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel _snake_case = logging.getLogger(__name__) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if os.path.exists(__a ): if os.path.exists(os.path.join(__a , "config.json" ) ) and os.path.isfile( os.path.join(__a , "config.json" ) ): os.remove(os.path.join(__a , "config.json" ) ) if os.path.exists(os.path.join(__a , "pytorch_model.bin" ) ) and os.path.isfile( os.path.join(__a , "pytorch_model.bin" ) ): os.remove(os.path.join(__a , "pytorch_model.bin" ) ) else: os.makedirs(__a ) model.save_pretrained(__a ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Optional[int] = 2 if unlogit: lowerCamelCase : List[str] = torch.pow(__a , __a ) lowerCamelCase : Union[str, Any] = p * torch.log(__a ) lowerCamelCase : int = 0 return -plogp.sum(dim=-1 ) def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' logger.info("lv, h >\t" + "\t".join(f"""{x + 1}""" for x in range(len(__a ) ) ) ) for row in range(len(__a ) ): if tensor.dtype != torch.long: logger.info(f"""layer {row + 1}:\t""" + "\t".join(f"""{x:.5f}""" for x in tensor[row].cpu().data ) ) else: logger.info(f"""layer {row + 1}:\t""" + "\t".join(f"""{x:d}""" for x in tensor[row].cpu().data ) ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Optional[Any] = model.config.num_hidden_layers, model.config.num_attention_heads lowerCamelCase : Optional[int] = torch.zeros(__a , __a ).to(args.device ) lowerCamelCase : Any = torch.zeros(__a , __a ).to(args.device ) if head_mask is None: lowerCamelCase : Union[str, Any] = torch.ones(__a , __a ).to(args.device ) head_mask.requires_grad_(requires_grad=__a ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: lowerCamelCase : List[str] = None lowerCamelCase : List[Any] = 0.0 lowerCamelCase : List[str] = 0.0 for step, inputs in enumerate(tqdm(__a , desc="Iteration" , disable=args.local_rank not in [-1, 0] ) ): lowerCamelCase : Optional[int] = tuple(t.to(args.device ) for t in inputs ) (lowerCamelCase ) : List[str] = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) lowerCamelCase : Any = model(__a , labels=__a , head_mask=__a ) # (loss), lm_logits, presents, (all hidden_states), (attentions) lowerCamelCase : Union[str, Any] = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__a ): lowerCamelCase : List[Any] = entropy(attn.detach() , __a ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__a ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: lowerCamelCase : str = 2 lowerCamelCase : Any = torch.pow(torch.pow(__a , __a ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1E-20 if not args.dont_normalize_global_importance: lowerCamelCase : str = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("Attention entropies" ) print_ad_tensor(__a ) if compute_importance: logger.info("Head importance scores" ) print_ad_tensor(__a ) logger.info("Head ranked by importance scores" ) lowerCamelCase : Optional[Any] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) lowerCamelCase : Any = torch.arange( head_importance.numel() , device=args.device ) lowerCamelCase : Union[str, Any] = head_ranks.view_as(__a ) print_ad_tensor(__a ) return attn_entropy, head_importance, total_loss def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : str = compute_heads_importance(__a , __a , __a , compute_entropy=__a ) lowerCamelCase : List[Any] = 1 / loss # instead of downsteam score use the LM loss logger.info("Pruning: original score: %f, threshold: %f" , __a , original_score * args.masking_threshold ) lowerCamelCase : Any = torch.ones_like(__a ) lowerCamelCase : Union[str, Any] = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) lowerCamelCase : List[str] = original_score while current_score >= original_score * args.masking_threshold: lowerCamelCase : str = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads lowerCamelCase : Tuple = float("Inf" ) lowerCamelCase : Any = head_importance.view(-1 ).sort()[1] if len(__a ) <= num_to_mask: print("BREAK BY num_to_mask" ) break # mask heads lowerCamelCase : int = current_heads_to_mask[:num_to_mask] logger.info("Heads to mask: %s" , str(current_heads_to_mask.tolist() ) ) lowerCamelCase : Tuple = new_head_mask.view(-1 ) lowerCamelCase : str = 0.0 lowerCamelCase : List[Any] = new_head_mask.view_as(__a ) lowerCamelCase : Any = new_head_mask.clone().detach() print_ad_tensor(__a ) # Compute metric and head importance again lowerCamelCase : Tuple = compute_heads_importance( __a , __a , __a , compute_entropy=__a , head_mask=__a ) lowerCamelCase : List[str] = 1 / loss logger.info( "Masking: current score: %f, remaining heads %d (%.1f percents)" , __a , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info("Final head mask" ) print_ad_tensor(__a ) np.save(os.path.join(args.output_dir , "head_mask.npy" ) , head_mask.detach().cpu().numpy() ) return head_mask def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[Any] = datetime.now() lowerCamelCase : Any = compute_heads_importance( __a , __a , __a , compute_entropy=__a , compute_importance=__a , head_mask=__a ) lowerCamelCase : List[Any] = 1 / loss lowerCamelCase : Union[str, Any] = datetime.now() - before_time lowerCamelCase : List[Any] = sum(p.numel() for p in model.parameters() ) lowerCamelCase : Tuple = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__a ) ) } for k, v in heads_to_prune.items(): if isinstance(__a , __a ): lowerCamelCase : List[str] = [ v, ] assert sum(len(__a ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__a ) lowerCamelCase : Optional[Any] = sum(p.numel() for p in model.parameters() ) lowerCamelCase : Union[str, Any] = datetime.now() lowerCamelCase : Any = compute_heads_importance( __a , __a , __a , compute_entropy=__a , compute_importance=__a , head_mask=__a , actually_pruned=__a , ) lowerCamelCase : int = 1 / loss lowerCamelCase : Any = datetime.now() - before_time logger.info( "Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)" , __a , __a , pruned_num_params / original_num_params * 100 , ) logger.info("Pruning: score with masking: %f score with pruning: %f" , __a , __a ) logger.info("Pruning: speed ratio (original timing / new timing): %f percents" , original_time / new_time * 100 ) save_model(__a , args.output_dir ) def lowercase_( ): '''simple docstring''' lowerCamelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--data_dir" , default=__a , type=__a , required=__a , help="The input data dir. Should contain the .tsv files (or other data files) for the task." , ) parser.add_argument( "--model_name_or_path" , default=__a , type=__a , required=__a , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--output_dir" , default=__a , type=__a , required=__a , help="The output directory where the model predictions and checkpoints will be written." , ) # Other parameters parser.add_argument( "--config_name" , default="" , type=__a , help="Pretrained config name or path if not the same as model_name_or_path" , ) parser.add_argument( "--tokenizer_name" , default="" , type=__a , help="Pretrained tokenizer name or path if not the same as model_name_or_path" , ) parser.add_argument( "--cache_dir" , default=__a , type=__a , help="Where do you want to store the pre-trained models downloaded from s3" , ) parser.add_argument( "--data_subset" , type=__a , default=-1 , help="If > 0: limit the data to a subset of data_subset instances." ) parser.add_argument( "--overwrite_output_dir" , action="store_true" , help="Whether to overwrite data in output directory" ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--dont_normalize_importance_by_layer" , action="store_true" , help="Don\'t normalize importance score by layers" ) parser.add_argument( "--dont_normalize_global_importance" , action="store_true" , help="Don\'t normalize all importance scores between 0 and 1" , ) parser.add_argument( "--try_masking" , action="store_true" , help="Whether to try to mask head until a threshold of accuracy." ) parser.add_argument( "--masking_threshold" , default=0.9 , type=__a , help="masking threshold in term of metrics (stop masking when metric < threshold * original metric value)." , ) parser.add_argument( "--masking_amount" , default=0.1 , type=__a , help="Amount to heads to masking at each masking step." ) parser.add_argument("--metric_name" , default="acc" , type=__a , help="Metric to use for head masking." ) parser.add_argument( "--max_seq_length" , default=128 , type=__a , help=( "The maximum total input sequence length after WordPiece tokenization. \n" "Sequences longer than this will be truncated, sequences shorter padded." ) , ) parser.add_argument("--batch_size" , default=1 , type=__a , help="Batch size." ) parser.add_argument("--seed" , type=__a , default=42 ) parser.add_argument("--local_rank" , type=__a , default=-1 , help="local_rank for distributed training on gpus" ) parser.add_argument("--no_cuda" , action="store_true" , help="Whether not to use CUDA when available" ) parser.add_argument("--server_ip" , type=__a , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=__a , default="" , help="Can be used for distant debugging." ) lowerCamelCase : str = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__a ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: lowerCamelCase : Optional[Any] = torch.device("cuda" if torch.cuda.is_available() and not args.no_cuda else "cpu" ) lowerCamelCase : List[str] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) lowerCamelCase : int = torch.device("cuda" , args.local_rank ) lowerCamelCase : Dict = 1 torch.distributed.init_process_group(backend="nccl" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) lowerCamelCase : List[Any] = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: lowerCamelCase : Any = nn.parallel.DistributedDataParallel( __a , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__a ) elif args.n_gpu > 1: lowerCamelCase : Union[str, Any] = nn.DataParallel(__a ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__a ) torch.save(__a , os.path.join(args.output_dir , "run_args.bin" ) ) logger.info("Training/evaluation parameters %s" , __a ) # Prepare dataset lowerCamelCase : Optional[Any] = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) lowerCamelCase : Optional[int] = (torch.from_numpy(__a ),) lowerCamelCase : Optional[int] = TensorDataset(*__a ) lowerCamelCase : List[str] = RandomSampler(__a ) lowerCamelCase : int = DataLoader(__a , sampler=__a , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__a , __a , __a ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: lowerCamelCase : Optional[int] = mask_heads(__a , __a , __a ) prune_heads(__a , __a , __a , __a ) if __name__ == "__main__": main()
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
def snake_case_ ( lowerCAmelCase_ : int = 100 ): __lowercase : Tuple = n * (n + 1) * (2 * n + 1) / 6 __lowercase : List[Any] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'''{solution() = }''')
355
from __future__ import annotations def snake_case_ ( lowerCAmelCase_ : str , lowerCAmelCase_ : str ): __lowercase : Any = get_failure_array(lowerCAmelCase_ ) # 2) Step through text searching for pattern __lowercase , __lowercase : Optional[int] = 0, 0 # index into text, pattern while i < len(lowerCAmelCase_ ): if pattern[j] == text[i]: if j == (len(lowerCAmelCase_ ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: __lowercase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( lowerCAmelCase_ : str ): __lowercase : List[Any] = [0] __lowercase : Optional[Any] = 0 __lowercase : List[Any] = 1 while j < len(lowerCAmelCase_ ): if pattern[i] == pattern[j]: i += 1 elif i > 0: __lowercase : List[str] = failure[i - 1] continue j += 1 failure.append(lowerCAmelCase_ ) return failure if __name__ == "__main__": # Test 1) lowerCamelCase : Dict = '''abc1abc12''' lowerCamelCase : Union[str, Any] = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' lowerCamelCase : Any = '''alskfjaldsk23adsfabcabc''' assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) lowerCamelCase : List[Any] = '''ABABX''' lowerCamelCase : List[Any] = '''ABABZABABYABABX''' assert kmp(pattern, text) # Test 3) lowerCamelCase : int = '''AAAB''' lowerCamelCase : Optional[int] = '''ABAAAAAB''' assert kmp(pattern, text) # Test 4) lowerCamelCase : Optional[Any] = '''abcdabcy''' lowerCamelCase : Any = '''abcxabcdabxabcdabcdabcy''' assert kmp(pattern, text) # Test 5) lowerCamelCase : Dict = '''aabaabaaa''' assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
306
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class _lowerCAmelCase : def __init__( self , _UpperCamelCase , _UpperCamelCase=13 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=99 , _UpperCamelCase=32 , _UpperCamelCase=5 , _UpperCamelCase=4 , _UpperCamelCase=37 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=512 , _UpperCamelCase=16 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=3 , _UpperCamelCase=4 , _UpperCamelCase=None , ) -> Dict: lowerCAmelCase_ = parent lowerCAmelCase_ = batch_size lowerCAmelCase_ = seq_length lowerCAmelCase_ = is_training lowerCAmelCase_ = use_token_type_ids lowerCAmelCase_ = use_labels lowerCAmelCase_ = vocab_size lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_act lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = type_vocab_size lowerCAmelCase_ = type_sequence_label_size lowerCAmelCase_ = initializer_range lowerCAmelCase_ = num_labels lowerCAmelCase_ = num_choices lowerCAmelCase_ = scope lowerCAmelCase_ = self.vocab_size - 1 def __a ( self ) -> List[str]: lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ = None if self.use_token_type_ids: lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase_ = None lowerCAmelCase_ = None lowerCAmelCase_ = None if self.use_labels: lowerCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) lowerCAmelCase_ = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , *_UpperCamelCase ) -> Tuple: lowerCAmelCase_ = OpenAIGPTModel(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase_ = model(_UpperCamelCase , token_type_ids=_UpperCamelCase , head_mask=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase , token_type_ids=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , *_UpperCamelCase ) -> Union[str, Any]: lowerCAmelCase_ = OpenAIGPTLMHeadModel(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase_ = model(_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , *_UpperCamelCase ) -> Any: lowerCAmelCase_ = OpenAIGPTDoubleHeadsModel(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase_ = model(_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , *_UpperCamelCase ) -> Optional[Any]: lowerCAmelCase_ = self.num_labels lowerCAmelCase_ = OpenAIGPTForSequenceClassification(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ = model(_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self ) -> List[str]: lowerCAmelCase_ = self.prepare_config_and_inputs() ( lowerCAmelCase_ ) = config_and_inputs lowerCAmelCase_ = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask, } return config, inputs_dict @require_torch class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): _lowercase =( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _lowercase =( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _lowercase =( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> List[Any]: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=False ) -> Union[str, Any]: lowerCAmelCase_ = super()._prepare_for_class(_UpperCamelCase , _UpperCamelCase , return_labels=_UpperCamelCase ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": lowerCAmelCase_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=_UpperCamelCase , ) lowerCAmelCase_ = inputs_dict["""labels"""] lowerCAmelCase_ = inputs_dict["""labels"""] lowerCAmelCase_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=_UpperCamelCase , ) lowerCAmelCase_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_UpperCamelCase ) return inputs_dict def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = OpenAIGPTModelTester(self ) lowerCAmelCase_ = ConfigTester(self , config_class=_UpperCamelCase , n_embd=37 ) def __a ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*_UpperCamelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*_UpperCamelCase ) def __a ( self ) -> Tuple: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*_UpperCamelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*_UpperCamelCase ) @slow def __a ( self ) -> List[str]: for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ = OpenAIGPTModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): @slow def __a ( self ) -> str: lowerCAmelCase_ = OpenAIGPTLMHeadModel.from_pretrained("openai-gpt" ) model.to(_UpperCamelCase ) lowerCAmelCase_ = torch.tensor([[481, 4_735, 544]] , dtype=torch.long , device=_UpperCamelCase ) # the president is lowerCAmelCase_ = [ 481, 4_735, 544, 246, 963, 870, 762, 239, 244, 40_477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the lowerCAmelCase_ = model.generate(_UpperCamelCase , do_sample=_UpperCamelCase ) self.assertListEqual(output_ids[0].tolist() , _UpperCamelCase )
231
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = "▁" lowerCamelCase_ = {"vocab_file": "sentencepiece.bpe.model"} lowerCamelCase_ = { "vocab_file": { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model" ), } } lowerCamelCase_ = { "xlm-roberta-base": 5_12, "xlm-roberta-large": 5_12, "xlm-roberta-large-finetuned-conll02-dutch": 5_12, "xlm-roberta-large-finetuned-conll02-spanish": 5_12, "xlm-roberta-large-finetuned-conll03-english": 5_12, "xlm-roberta-large-finetuned-conll03-german": 5_12, } class lowercase_ ( A ): """simple docstring""" lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = ['''input_ids''', '''attention_mask'''] def __init__( self : str , __lowerCamelCase : int , __lowerCamelCase : Optional[int]="<s>" , __lowerCamelCase : Optional[int]="</s>" , __lowerCamelCase : List[Any]="</s>" , __lowerCamelCase : str="<s>" , __lowerCamelCase : int="<unk>" , __lowerCamelCase : Dict="<pad>" , __lowerCamelCase : Tuple="<mask>" , __lowerCamelCase : Optional[Dict[str, Any]] = None , **__lowerCamelCase : Optional[Any] , ): """simple docstring""" _SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else mask_token _SCREAMING_SNAKE_CASE = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , sep_token=__lowerCamelCase , cls_token=__lowerCamelCase , pad_token=__lowerCamelCase , mask_token=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCamelCase , ) _SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _SCREAMING_SNAKE_CASE = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = len(self.sp_model ) + self.fairseq_offset _SCREAMING_SNAKE_CASE = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.__dict__.copy() _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() return state def __setstate__( self : str , __lowerCamelCase : List[str] ): """simple docstring""" _SCREAMING_SNAKE_CASE = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def lowerCAmelCase_ ( self : Any , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] _SCREAMING_SNAKE_CASE = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCAmelCase_ ( self : int , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) if token_ids_a is None: return [1] + ([0] * len(__lowerCamelCase )) + [1] return [1] + ([0] * len(__lowerCamelCase )) + [1, 1] + ([0] * len(__lowerCamelCase )) + [1] def lowerCAmelCase_ ( self : Tuple , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): """simple docstring""" _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = {self.convert_ids_to_tokens(__lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCAmelCase_ ( self : List[str] , __lowerCamelCase : str ): """simple docstring""" return self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase ) def lowerCAmelCase_ ( self : Any , __lowerCamelCase : str ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _SCREAMING_SNAKE_CASE = self.sp_model.PieceToId(__lowerCamelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCAmelCase_ ( self : Tuple , __lowerCamelCase : Union[str, Any] ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowerCAmelCase_ ( self : Optional[Any] , __lowerCamelCase : List[str] ): """simple docstring""" _SCREAMING_SNAKE_CASE = "".join(__lowerCamelCase ).replace(__lowerCamelCase , " " ).strip() return out_string def lowerCAmelCase_ ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ): """simple docstring""" if not os.path.isdir(__lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _SCREAMING_SNAKE_CASE = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCamelCase , "wb" ) as fi: _SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase ) return (out_vocab_file,)
358
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase_ = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
111
0
"""simple docstring""" def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: while a != 0: snake_case_ = b % a, a return b def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: if gcd(UpperCAmelCase_ , UpperCAmelCase_ ) != 1: snake_case_ = f"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(UpperCAmelCase_ ) snake_case_ = 1, 0, a snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
347
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : List[str]= logging.get_logger(__name__) _a : Optional[int]= { "weiweishi/roc-bert-base-zh": "https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json", } class UpperCamelCase ( lowercase ): UpperCAmelCase : Union[str, Any] = """roc_bert""" def __init__(self : Dict , _A : str=3_05_22 , _A : List[str]=7_68 , _A : int=12 , _A : int=12 , _A : Any=30_72 , _A : List[Any]="gelu" , _A : str=0.1 , _A : Tuple=0.1 , _A : Optional[Any]=5_12 , _A : Optional[Any]=2 , _A : Dict=0.02 , _A : Tuple=1E-12 , _A : Any=True , _A : List[str]=0 , _A : Tuple="absolute" , _A : List[str]=None , _A : Union[str, Any]=True , _A : Tuple=True , _A : Dict=7_68 , _A : str=9_10 , _A : List[str]=5_12 , _A : str=2_48_58 , _A : Tuple=True , **_A : Dict , ) -> List[Any]: __snake_case : Tuple = vocab_size __snake_case : List[Any] = max_position_embeddings __snake_case : Tuple = hidden_size __snake_case : Optional[Any] = num_hidden_layers __snake_case : Tuple = num_attention_heads __snake_case : Any = intermediate_size __snake_case : List[Any] = hidden_act __snake_case : Any = hidden_dropout_prob __snake_case : Union[str, Any] = attention_probs_dropout_prob __snake_case : List[str] = initializer_range __snake_case : Optional[Any] = type_vocab_size __snake_case : str = layer_norm_eps __snake_case : Dict = use_cache __snake_case : int = enable_pronunciation __snake_case : Dict = enable_shape __snake_case : Any = pronunciation_embed_dim __snake_case : Union[str, Any] = pronunciation_vocab_size __snake_case : List[Any] = shape_embed_dim __snake_case : List[Any] = shape_vocab_size __snake_case : List[Any] = concat_input __snake_case : str = position_embedding_type __snake_case : List[Any] = classifier_dropout super().__init__(pad_token_id=_A , **_A)
172
0
def __UpperCAmelCase ( __a : int = 1_000_000 ) -> int: """simple docstring""" _a : List[str] = [i - 1 for i in range(limit + 1 )] for i in range(2 ,limit + 1 ): if phi[i] == i - 1: for j in range(2 * i ,limit + 1 ,__a ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
15
from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch a__ = logging.get_logger(__name__) @add_end_docstrings( __lowercase , r"\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n " , ) class UpperCAmelCase_ ( __lowercase ): """simple docstring""" def __lowercase ( self , _a ) -> np.ndarray: if self.framework == "tf": _a : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": _a : Tuple = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=_a ) else: raise ValueError('''Unsupported framework''' ) return masked_index def __lowercase ( self , _a ) -> np.ndarray: _a : int = self.get_masked_index(_a ) _a : Tuple = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( '''fill-mask''' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def __lowercase ( self , _a ) -> Optional[int]: if isinstance(_a , _a ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['''input_ids'''][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(_a ) def __lowercase ( self , _a , _a=None , **_a ) -> Dict[str, GenericTensor]: if return_tensors is None: _a : Union[str, Any] = self.framework _a : str = self.tokenizer(_a , return_tensors=_a ) self.ensure_exactly_one_mask_token(_a ) return model_inputs def __lowercase ( self , _a ) -> Optional[Any]: _a : List[str] = self.model(**_a ) _a : Any = model_inputs['''input_ids'''] return model_outputs def __lowercase ( self , _a , _a=5 , _a=None ) -> str: # Cap top_k if there are targets if target_ids is not None and target_ids.shape[0] < top_k: _a : List[Any] = target_ids.shape[0] _a : Any = model_outputs['''input_ids'''][0] _a : List[str] = model_outputs['''logits'''] if self.framework == "tf": _a : Tuple = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] _a : List[str] = outputs.numpy() _a : Dict = outputs[0, masked_index, :] _a : str = stable_softmax(_a , axis=-1 ) if target_ids is not None: _a : Any = tf.gather_nd(tf.squeeze(_a , 0 ) , target_ids.reshape(-1 , 1 ) ) _a : Union[str, Any] = tf.expand_dims(_a , 0 ) _a : Optional[int] = tf.math.top_k(_a , k=_a ) _a , _a : Optional[Any] = topk.values.numpy(), topk.indices.numpy() else: _a : Optional[Any] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=_a ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample _a : List[str] = outputs[0, masked_index, :] _a : List[Any] = logits.softmax(dim=-1 ) if target_ids is not None: _a : List[Any] = probs[..., target_ids] _a , _a : Optional[Any] = probs.topk(_a ) _a : Dict = [] _a : List[Any] = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): _a : Optional[Any] = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place _a : Optional[int] = input_ids.numpy().copy() if target_ids is not None: _a : Tuple = target_ids[p].tolist() _a : List[str] = p # Filter padding out: _a : List[Any] = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back _a : List[str] = self.tokenizer.decode(_a , skip_special_tokens=_a ) _a : List[Any] = {'''score''': v, '''token''': p, '''token_str''': self.tokenizer.decode([p] ), '''sequence''': sequence} row.append(_a ) result.append(_a ) if single_mask: return result[0] return result def __lowercase ( self , _a , _a=None ) -> Dict: if isinstance(_a , _a ): _a : Tuple = [targets] try: _a : int = self.tokenizer.get_vocab() except Exception: _a : Any = {} _a : List[Any] = [] for target in targets: _a : List[Any] = vocab.get(_a , _a ) if id_ is None: _a : Tuple = self.tokenizer( _a , add_special_tokens=_a , return_attention_mask=_a , return_token_type_ids=_a , max_length=1 , truncation=_a , )['''input_ids'''] if len(_a ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ '''We cannot replace it with anything meaningful, ignoring it''' ) continue _a : Tuple = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) _a : List[str] = list(set(_a ) ) if len(_a ) == 0: raise ValueError('''At least one target must be provided when passed.''' ) _a : int = np.array(_a ) return target_ids def __lowercase ( self , _a=None , _a=None ) -> Tuple: _a : str = {} if targets is not None: _a : List[Any] = self.get_target_ids(_a , _a ) _a : Optional[Any] = target_ids if top_k is not None: _a : Union[str, Any] = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( '''fill-mask''' , self.model.base_model_prefix , '''The tokenizer does not define a `mask_token`.''' ) return {}, {}, postprocess_params def __call__( self , _a , *_a , **_a ) -> int: _a : Optional[Any] = super().__call__(_a , **_a ) if isinstance(_a , _a ) and len(_a ) == 1: return outputs[0] return outputs
15
1
"""simple docstring""" from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig a : Any = logging.get_logger(__name__) # General docstring a : Optional[Any] = '''RegNetConfig''' # Base docstring a : str = '''facebook/regnet-y-040''' a : Optional[Any] = [1, 1088, 7, 7] # Image classification docstring a : Tuple = '''facebook/regnet-y-040''' a : Dict = '''tabby, tabby cat''' a : int = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = "relu" , ) -> Optional[Any]: super().__init__() a : Optional[Any] = nn.Convad( lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=lowerCAmelCase__ , stride=lowerCAmelCase__ , padding=kernel_size // 2 , groups=lowerCAmelCase__ , bias=lowerCAmelCase__ , ) a : int = nn.BatchNormad(lowerCAmelCase__ ) a : List[str] = ACTaFN[activation] if activation is not None else nn.Identity() def __a ( self , lowerCAmelCase__ ) -> int: a : List[str] = self.convolution(lowerCAmelCase__ ) a : Any = self.normalization(lowerCAmelCase__ ) a : Optional[Any] = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ ) -> int: super().__init__() a : Union[str, Any] = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) a : List[str] = config.num_channels def __a ( self , lowerCAmelCase__ ) -> Tuple: a : Union[str, Any] = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) a : Any = self.embedder(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 2 ) -> List[Any]: super().__init__() a : List[Any] = nn.Convad(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 , stride=lowerCAmelCase__ , bias=lowerCAmelCase__ ) a : Union[str, Any] = nn.BatchNormad(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> Tensor: a : List[Any] = self.convolution(lowerCAmelCase__ ) a : List[Any] = self.normalization(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: super().__init__() a : int = nn.AdaptiveAvgPoolad((1, 1) ) a : Dict = nn.Sequential( nn.Convad(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 ) , nn.ReLU() , nn.Convad(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 ) , nn.Sigmoid() , ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: # b c h w -> b c 1 1 a : int = self.pooler(lowerCAmelCase__ ) a : Optional[Any] = self.attention(lowerCAmelCase__ ) a : Any = hidden_state * attention return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 ) -> List[str]: super().__init__() a : str = in_channels != out_channels or stride != 1 a : str = max(1 , out_channels // config.groups_width ) a : int = ( RegNetShortCut(lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ ) if should_apply_shortcut else nn.Identity() ) a : Tuple = nn.Sequential( RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ , groups=lowerCAmelCase__ , activation=config.hidden_act ) , RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 , activation=lowerCAmelCase__ ) , ) a : List[Any] = ACTaFN[config.hidden_act] def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: a : Any = hidden_state a : int = self.layer(lowerCAmelCase__ ) a : Union[str, Any] = self.shortcut(lowerCAmelCase__ ) hidden_state += residual a : Tuple = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 ) -> str: super().__init__() a : Optional[int] = in_channels != out_channels or stride != 1 a : str = max(1 , out_channels // config.groups_width ) a : int = ( RegNetShortCut(lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ ) if should_apply_shortcut else nn.Identity() ) a : int = nn.Sequential( RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ , groups=lowerCAmelCase__ , activation=config.hidden_act ) , RegNetSELayer(lowerCAmelCase__ , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(lowerCAmelCase__ , lowerCAmelCase__ , kernel_size=1 , activation=lowerCAmelCase__ ) , ) a : Optional[Any] = ACTaFN[config.hidden_act] def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: a : str = hidden_state a : int = self.layer(lowerCAmelCase__ ) a : Any = self.shortcut(lowerCAmelCase__ ) hidden_state += residual a : int = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 2 , ) -> int: super().__init__() a : Optional[int] = RegNetXLayer if config.layer_type == "x" else RegNetYLayer a : Optional[int] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ , ) , *[layer(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for _ in range(depth - 1 )] , ) def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: a : Optional[Any] = self.layers(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ ) -> str: super().__init__() a : Any = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( lowerCAmelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) a : Any = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(lowerCAmelCase__ , config.depths[1:] ): self.stages.append(RegNetStage(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , depth=lowerCAmelCase__ ) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False , lowerCAmelCase__ = True ) -> BaseModelOutputWithNoAttention: a : Any = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: a : Any = hidden_states + (hidden_state,) a : Any = stage_module(lowerCAmelCase__ ) if output_hidden_states: a : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=lowerCAmelCase__ , hidden_states=lowerCAmelCase__ ) class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =RegNetConfig lowerCamelCase : Dict ="""regnet""" lowerCamelCase : Tuple ="""pixel_values""" lowerCamelCase : List[Any] =True def __a ( self , lowerCAmelCase__ ) -> List[Any]: if isinstance(lowerCAmelCase__ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu" ) elif isinstance(lowerCAmelCase__ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=False ) -> List[Any]: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = value a : str = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' a : Optional[int] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare RegNet model outputting raw features without any specific head on top.""" , a__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ ) -> int: super().__init__(lowerCAmelCase__ ) a : Tuple = config a : Optional[int] = RegNetEmbeddings(lowerCAmelCase__ ) a : List[str] = RegNetEncoder(lowerCAmelCase__ ) a : Tuple = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None ) -> BaseModelOutputWithPoolingAndNoAttention: a : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : str = return_dict if return_dict is not None else self.config.use_return_dict a : Tuple = self.embedder(lowerCAmelCase__ ) a : Any = self.encoder( lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) a : Any = encoder_outputs[0] a : Union[str, Any] = self.pooler(lowerCAmelCase__ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCAmelCase__ , pooler_output=lowerCAmelCase__ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( """ RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , a__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ ) -> str: super().__init__(lowerCAmelCase__ ) a : List[str] = config.num_labels a : Dict = RegNetModel(lowerCAmelCase__ ) # classification head a : Optional[Any] = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __a ( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , ) -> ImageClassifierOutputWithNoAttention: a : Dict = return_dict if return_dict is not None else self.config.use_return_dict a : List[Any] = self.regnet(lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) a : Any = outputs.pooler_output if return_dict else outputs[1] a : List[str] = self.classifier(lowerCAmelCase__ ) a : str = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: a : Optional[int] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): a : Optional[Any] = "single_label_classification" else: a : str = "multi_label_classification" if self.config.problem_type == "regression": a : Optional[Any] = MSELoss() if self.num_labels == 1: a : List[Any] = loss_fct(logits.squeeze() , labels.squeeze() ) else: a : str = loss_fct(lowerCAmelCase__ , lowerCAmelCase__ ) elif self.config.problem_type == "single_label_classification": a : Union[str, Any] = CrossEntropyLoss() a : Tuple = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": a : Dict = BCEWithLogitsLoss() a : List[Any] = loss_fct(lowerCAmelCase__ , lowerCAmelCase__ ) if not return_dict: a : int = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowerCAmelCase__ , logits=lowerCAmelCase__ , hidden_states=outputs.hidden_states )
105
"""simple docstring""" import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __UpperCamelCase : @staticmethod def __a ( *lowerCAmelCase__ , **lowerCAmelCase__ ) -> str: pass def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->Dict: '''simple docstring''' return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. a : Optional[Any] = ( '''https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png''' ) @is_pipeline_test @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): lowerCamelCase : Union[str, Any] =MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : Tuple = pipeline( "document-question-answering" , model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) a : Optional[int] = INVOICE_URL a : str = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , "" ) ) ) a : Union[str, Any] = "What is the placebo?" a : Dict = [ { "image": load_image(lowerCAmelCase__ ), "question": question, }, { "image": image, "question": question, }, { "image": image, "question": question, "word_boxes": word_boxes, }, ] return dqa_pipeline, examples def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: a : Tuple = dqa_pipeline(lowerCAmelCase__ , top_k=2 ) self.assertEqual( lowerCAmelCase__ , [ [ {"score": ANY(lowerCAmelCase__ ), "answer": ANY(lowerCAmelCase__ ), "start": ANY(lowerCAmelCase__ ), "end": ANY(lowerCAmelCase__ )}, {"score": ANY(lowerCAmelCase__ ), "answer": ANY(lowerCAmelCase__ ), "start": ANY(lowerCAmelCase__ ), "end": ANY(lowerCAmelCase__ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def __a ( self ) -> List[Any]: a : List[Any] = pipeline("document-question-answering" , model="hf-internal-testing/tiny-random-layoutlmv2" ) a : Dict = INVOICE_URL a : List[str] = "How many cats are there?" a : Tuple = [ {"score": 0.0_001, "answer": "oy 2312/2019", "start": 38, "end": 39}, {"score": 0.0_001, "answer": "oy 2312/2019 DUE", "start": 38, "end": 40}, ] a : Optional[int] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) a : Optional[int] = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably a : List[Any] = "./tests/fixtures/tests_samples/COCO/000000039769.png" a : Any = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) # We can optionnally pass directly the words and bounding boxes a : Optional[int] = "./tests/fixtures/tests_samples/COCO/000000039769.png" a : Tuple = [] a : Optional[int] = [] a : List[str] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , words=lowerCAmelCase__ , boxes=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def __a ( self ) -> Tuple: a : int = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , ) a : List[str] = INVOICE_URL a : List[Any] = "What is the invoice number?" a : int = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_009, "answer": "us-001", "start": 16, "end": 16}, ] , ) a : str = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_009, "answer": "us-001", "start": 16, "end": 16}, ] , ) a : Any = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"score": 0.9_944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_009, "answer": "us-001", "start": 16, "end": 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def __a ( self ) -> Optional[int]: a : List[str] = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , max_seq_len=50 , ) a : Optional[Any] = INVOICE_URL a : Tuple = "What is the invoice number?" a : List[Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9_948, "answer": "us-001", "start": 16, "end": 16}, ] , ) a : str = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9_948, "answer": "us-001", "start": 16, "end": 16}, ] , ) a : Tuple = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"score": 0.9_974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9_948, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def __a ( self ) -> str: a : Optional[int] = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=lowerCAmelCase__ ) a : int = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=lowerCAmelCase__ , revision="3dc6de3" , ) a : List[Any] = INVOICE_URL a : Tuple = "What is the invoice number?" a : List[str] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.4_251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) a : Dict = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.4_251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) a : List[Any] = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"score": 0.4_251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_819, "answer": "1110212019", "start": 23, "end": 23}, ] ] * 2 , ) a : Dict = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , "" ) ) ) # This model should also work if `image` is set to None a : Optional[Any] = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.4_251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0_819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def __a ( self ) -> Tuple: a : int = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=lowerCAmelCase__ ) a : Tuple = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=lowerCAmelCase__ , revision="3dc6de3" , max_seq_len=50 , ) a : List[str] = INVOICE_URL a : Union[str, Any] = "What is the invoice number?" a : List[Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9_998, "answer": "us-001", "start": 16, "end": 16}, ] , ) a : List[str] = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"score": 0.9_999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9_998, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) a : List[Any] = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , "" ) ) ) # This model should also work if `image` is set to None a : Any = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"score": 0.9_999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9_998, "answer": "us-001", "start": 16, "end": 16}, ] , ) @slow @require_torch def __a ( self ) -> int: a : Tuple = pipeline( "document-question-answering" , model="naver-clova-ix/donut-base-finetuned-docvqa" , tokenizer=AutoTokenizer.from_pretrained("naver-clova-ix/donut-base-finetuned-docvqa" ) , feature_extractor="naver-clova-ix/donut-base-finetuned-docvqa" , ) a : Optional[Any] = INVOICE_URL a : Tuple = "What is the invoice number?" a : List[Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , [{"answer": "us-001"}] ) @require_tf @unittest.skip("Document question answering not implemented in TF" ) def __a ( self ) -> int: pass
105
1
"""simple docstring""" def __snake_case ( _lowerCAmelCase : int ) -> int: if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return bin(_lowerCAmelCase ).count("1" ) if __name__ == "__main__": import doctest doctest.testmod()
369
from math import pi, sqrt def __snake_case ( _lowerCAmelCase : float ) -> float: if num <= 0: raise ValueError("math domain error" ) if num > 1_71.5: raise OverflowError("math range error" ) elif num - int(_lowerCAmelCase ) not in (0, 0.5): raise NotImplementedError("num must be an integer or a half-integer" ) elif num == 0.5: return sqrt(_lowerCAmelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def __snake_case ( ) -> None: assert gamma(0.5 ) == sqrt(_lowerCAmelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() _lowerCAmelCase : List[str] = 1.0 while num: _lowerCAmelCase : List[str] = float(input('''Gamma of: ''')) print(F'''gamma({num}) = {gamma(num)}''') print('''\nEnter 0 to exit...''')
70
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __UpperCamelCase ( lowerCamelCase__ ): lowercase : Optional[int] ='facebook/bart-large-mnli' lowercase : List[Any] =( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) lowercase : Optional[int] ='text_classifier' lowercase : List[Any] =AutoTokenizer lowercase : Optional[int] =AutoModelForSequenceClassification lowercase : Union[str, Any] =['text', ['text']] lowercase : int =['text'] def lowercase__ ( self ): """simple docstring""" super().setup() lowerCamelCase_ =self.model.config lowerCamelCase_ =-1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): lowerCamelCase_ =int(lowerCAmelCase ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =labels return self.pre_processor( [text] * len(lowerCAmelCase ), [f'''This example is {label}''' for label in labels], return_tensors='''pt''', padding='''max_length''', ) def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =outputs.logits lowerCamelCase_ =torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
75
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' def __init__( self , *__A , **__A ): """simple docstring""" warnings.warn( "The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DPTImageProcessor instead." , __A , ) super().__init__(*__A , **__A )
283
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig UpperCAmelCase_ = { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/config.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/config.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/config.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/config.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/config.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/config.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json', } class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : Any = """albert""" def __init__( self : List[str] , _UpperCAmelCase : Tuple=3_00_00 , _UpperCAmelCase : Optional[int]=1_28 , _UpperCAmelCase : Dict=40_96 , _UpperCAmelCase : List[str]=12 , _UpperCAmelCase : Union[str, Any]=1 , _UpperCAmelCase : Optional[Any]=64 , _UpperCAmelCase : List[Any]=1_63_84 , _UpperCAmelCase : Tuple=1 , _UpperCAmelCase : Union[str, Any]="gelu_new" , _UpperCAmelCase : str=0 , _UpperCAmelCase : str=0 , _UpperCAmelCase : Tuple=5_12 , _UpperCAmelCase : Dict=2 , _UpperCAmelCase : Tuple=0.02 , _UpperCAmelCase : List[Any]=1E-12 , _UpperCAmelCase : Any=0.1 , _UpperCAmelCase : Dict="absolute" , _UpperCAmelCase : int=0 , _UpperCAmelCase : Optional[Any]=2 , _UpperCAmelCase : List[Any]=3 , **_UpperCAmelCase : Any , ): """simple docstring""" super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = embedding_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_hidden_groups UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = inner_group_num UpperCAmelCase__ = hidden_act UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = classifier_dropout_prob UpperCAmelCase__ = position_embedding_type class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' @property def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
353
'''simple docstring''' import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = GPTSanJapaneseTokenizer lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Union[str, Any] = {"""do_clean_text""": False, """add_prefix_space""": False} def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" super().setUp() # fmt: off UpperCAmelCase__ = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase__ = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 UpperCAmelCase__ = {"""unk_token""": """<unk>"""} UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , **_UpperCAmelCase : Optional[int] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : Any ): """simple docstring""" UpperCAmelCase__ = """こんにちは、世界。 \nこんばんは、㔺界。😀""" UpperCAmelCase__ = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def SCREAMING_SNAKE_CASE__ ( self : List[str] , _UpperCAmelCase : str ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.get_input_output_texts(_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.decode(_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase ) return text, ids def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" pass # TODO add if relevant def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" pass # TODO add if relevant def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" pass # TODO add if relevant def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" UpperCAmelCase__ = self.get_tokenizer() # Testing tokenization UpperCAmelCase__ = """こんにちは、世界。 こんばんは、㔺界。""" UpperCAmelCase__ = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] UpperCAmelCase__ = tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) # Testing conversion to ids without special tokens UpperCAmelCase__ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) # Testing conversion to ids with special tokens UpperCAmelCase__ = tokens + [tokenizer.unk_token] UpperCAmelCase__ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" UpperCAmelCase__ = self.get_tokenizer() # Testing tokenization UpperCAmelCase__ = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" UpperCAmelCase__ = """こんにちは、、、、世界。こんばんは、、、、世界。""" UpperCAmelCase__ = tokenizer.encode(_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" UpperCAmelCase__ = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization UpperCAmelCase__ = """こんにちは、世界。""" UpperCAmelCase__ = """こんばんは、㔺界。😀""" UpperCAmelCase__ = """こんにちは、世界。こんばんは、世界。😀""" UpperCAmelCase__ = tokenizer.encode(prefix_text + input_text ) UpperCAmelCase__ = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) UpperCAmelCase__ = tokenizer.encode(_UpperCAmelCase , prefix_text=_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.decode(_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.decode(_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" UpperCAmelCase__ = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization UpperCAmelCase__ = """こんにちは、世界。""" UpperCAmelCase__ = """こんばんは、㔺界。😀""" UpperCAmelCase__ = len(tokenizer.encode(_UpperCAmelCase ) ) - 2 UpperCAmelCase__ = len(tokenizer.encode(_UpperCAmelCase ) ) - 2 UpperCAmelCase__ = [1] + [0] * (len_prefix + len_text + 1) UpperCAmelCase__ = [1] * (len_prefix + len_text + 1) + [0] UpperCAmelCase__ = [1] + [1] * (len_prefix) + [0] * (len_text + 1) UpperCAmelCase__ = tokenizer(prefix_text + input_text ).token_type_ids UpperCAmelCase__ = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids UpperCAmelCase__ = tokenizer(_UpperCAmelCase , prefix_text=_UpperCAmelCase ).token_type_ids self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" UpperCAmelCase__ = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) UpperCAmelCase__ = tokenizer.encode("""あンいワ""" ) UpperCAmelCase__ = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) UpperCAmelCase__ = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(_UpperCAmelCase ) , tokenizer.decode(_UpperCAmelCase ) ) self.assertEqual(tokenizer.decode(_UpperCAmelCase ) , tokenizer.decode(_UpperCAmelCase ) ) self.assertNotEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertNotEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" UpperCAmelCase__ = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) UpperCAmelCase__ = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] UpperCAmelCase__ = tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.batch_encode_plus(_UpperCAmelCase , padding=_UpperCAmelCase ) # fmt: off UpperCAmelCase__ = [[3_59_93, 86_40, 2_59_48, 3_59_98, 3_06_47, 3_56_75, 3_59_99, 3_59_99], [3_59_93, 1_03_82, 98_68, 3_59_98, 3_06_46, 94_59, 3_06_46, 3_56_75]] UpperCAmelCase__ = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] UpperCAmelCase__ = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , _UpperCAmelCase ) self.assertListEqual(x_token.token_type_ids , _UpperCAmelCase ) self.assertListEqual(x_token.attention_mask , _UpperCAmelCase ) self.assertListEqual(x_token_a.input_ids , _UpperCAmelCase ) self.assertListEqual(x_token_a.token_type_ids , _UpperCAmelCase ) self.assertListEqual(x_token_a.attention_mask , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" pass
61
0
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __A = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): """simple docstring""" snake_case_ = XLNetTokenizer snake_case_ = XLNetTokenizerFast snake_case_ = True snake_case_ = True def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLNetTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = '<s>' __lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<eod>' ) self.assertEqual(len(lowerCamelCase__ ) , 1_006 ) def lowercase_ ( self ) -> Tuple: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowercase_ ( self ) -> List[str]: '''simple docstring''' __lowerCamelCase = XLNetTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) __lowerCamelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [285, 46, 10, 170, 382] ) __lowerCamelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = XLNetTokenizer(lowerCamelCase__ , do_lower_case=lowerCamelCase__ ) __lowerCamelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '', 'i', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['▁he', 'll', 'o'] ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = XLNetTokenizer(lowerCamelCase__ , do_lower_case=lowerCamelCase__ ) __lowerCamelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) @slow def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = XLNetTokenizer.from_pretrained('xlnet-base-cased' ) __lowerCamelCase = tokenizer.encode('sequence builders' , add_special_tokens=lowerCamelCase__ ) __lowerCamelCase = tokenizer.encode('multi-sequence build' , add_special_tokens=lowerCamelCase__ ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ , lowerCamelCase__ ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def lowercase_ ( self ) -> str: '''simple docstring''' # fmt: off __lowerCamelCase = {'input_ids': [[17, 21_442, 270, 17, 10, 14_645, 318, 34, 17, 4_546, 3_145, 787, 13, 7_752, 22_018, 23, 21, 17, 4_546, 3_145, 787, 13, 3_352, 14_431, 13, 5_500, 11, 1_176, 580, 13, 16_819, 4_797, 23, 17, 10, 17_135, 658, 19, 457, 7_932, 13, 184, 19, 3_154, 17_135, 6_468, 19, 1_404, 12_269, 19, 4_229, 5_356, 16_264, 46, 19, 17, 20_545, 10_395, 9, 9, 9, 11, 28, 6_421, 9_531, 20_729, 17, 10, 353, 17_022, 11, 21, 6_421, 9_531, 16_949, 17, 10, 11_509, 753, 11, 33, 95, 2_421, 7_385, 956, 14_431, 2_626, 25, 842, 7_385, 4_836, 21, 1_429, 2_272, 9_855, 3_120, 161, 24_738, 19, 13_203, 658, 218, 787, 21, 430, 18_482, 847, 2_637, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 322, 22_178, 27, 1_064, 22, 956, 13, 11_101, 1_429, 5_854, 24_313, 18_953, 40, 422, 24_366, 68, 1_758, 37, 10_483, 14_257, 31, 207, 263, 21, 203, 3_773, 25, 71, 9_735, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 2_049, 3_442, 17, 13_894, 3_380, 23, 95, 18, 17_634, 2_288, 9, 4, 3]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='xlnet-base-cased' , revision='c841166438c31ec7ca9a106dee7bb312b73ae511' , )
90
"""simple docstring""" lowerCAmelCase__ = { '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } def snake_case_ ( A_ : dict, A_ : int, A_ : int ): '''simple docstring''' _lowerCamelCase : List[str] = set() # keep track of all the paths to be checked _lowerCamelCase : str = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue _lowerCamelCase : str = queue.pop(0 ) # get the last node from the path _lowerCamelCase : List[Any] = path[-1] if node not in explored: _lowerCamelCase : Union[str, Any] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: _lowerCamelCase : Union[str, Any] = list(A_ ) new_path.append(A_ ) queue.append(A_ ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(A_ ) # in case there's no path between the 2 nodes return [] def snake_case_ ( A_ : dict, A_ : int, A_ : Dict ): '''simple docstring''' if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 _lowerCamelCase : Optional[int] = [start] _lowerCamelCase : int = set(A_ ) # Keep tab on distances from `start` node. _lowerCamelCase : int = {start: 0, target: -1} while queue: _lowerCamelCase : Optional[Any] = queue.pop(0 ) if node == target: _lowerCamelCase : Any = ( dist[node] if dist[target] == -1 else min(dist[target], dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(A_ ) queue.append(A_ ) _lowerCamelCase : Any = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, '''G''', '''D''')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, '''G''', '''D''')) # returns 4
72
0
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowerCAmelCase ( UpperCamelCase__ , unittest.TestCase): _lowercase : List[Any] = LxmertTokenizer _lowercase : Dict = LxmertTokenizerFast _lowercase : int = True _lowercase : Optional[Any] = True def _lowercase ( self ) -> str: '''simple docstring''' super().setUp() a__ : List[Any] =[ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] a__ : Optional[Any] =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def _lowercase ( self , lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' a__ : Optional[Any] ="UNwant\u00E9d,running" a__ : List[str] ="unwanted, running" return input_text, output_text def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : List[str] =self.tokenizer_class(self.vocab_file ) a__ : Dict =tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(lowerCAmelCase__ , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [7, 4, 5, 1_0, 8, 9] ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' if not self.test_rust_tokenizer: return a__ : List[Any] =self.get_tokenizer() a__ : int =self.get_rust_tokenizer() a__ : Union[str, Any] ="I was born in 92000, and this is falsé." a__ : List[Any] =tokenizer.tokenize(lowerCAmelCase__ ) a__ : Dict =rust_tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) a__ : List[Any] =tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) a__ : int =rust_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) a__ : Optional[int] =self.get_rust_tokenizer() a__ : List[Any] =tokenizer.encode(lowerCAmelCase__ ) a__ : List[str] =rust_tokenizer.encode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ )
148
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. UpperCAmelCase : List[Any] = {"""LayoutLMv2Config""", """LayoutLMv3Config"""} @is_pipeline_test class __lowerCAmelCase ( unittest.TestCase): _lowercase : str = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowercase : Tuple = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: _lowercase : Dict = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: _lowercase : int = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : str =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="pt" ) a__ : Any =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) a__ : Any =text_classifier("This is great !" , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}] ) a__ : Tuple =text_classifier(["This is great !", "This is bad"] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], ] , ) a__ : List[Any] =text_classifier("This is great !" , top_k=1 ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) # Legacy behavior a__ : Any =text_classifier("This is great !" , return_all_scores=lowerCAmelCase__ ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) a__ : List[str] =text_classifier("This is great !" , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [[{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}]] ) a__ : Optional[int] =text_classifier(["This is great !", "Something else"] , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], ] , ) a__ : int =text_classifier(["This is great !", "Something else"] , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ {"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_0", "score": 0.5_04}, ] , ) @require_torch def _lowercase ( self ) -> List[Any]: '''simple docstring''' import torch a__ : Dict =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="pt" , device=torch.device("cpu" ) , ) a__ : Optional[Any] =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) @require_tf def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : Optional[int] =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="tf" ) a__ : Optional[Any] =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) @slow @require_torch def _lowercase ( self ) -> str: '''simple docstring''' a__ : Union[str, Any] =pipeline("text-classification" ) a__ : Union[str, Any] =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 1.0}] ) a__ : Optional[Any] =text_classifier("This is bad !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "NEGATIVE", "score": 1.0}] ) a__ : Dict =text_classifier("Birds are a type of animal" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 0.9_88}] ) @slow @require_tf def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Tuple =pipeline("text-classification" , framework="tf" ) a__ : str =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 1.0}] ) a__ : str =text_classifier("This is bad !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "NEGATIVE", "score": 1.0}] ) a__ : Dict =text_classifier("Birds are a type of animal" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 0.9_88}] ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' a__ : List[Any] =TextClassificationPipeline(model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) return text_classifier, ["HuggingFace is in", "This is another test"] def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: '''simple docstring''' a__ : Tuple =text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 a__ : List[Any] ="HuggingFace is in" a__ : int =text_classifier(lowerCAmelCase__ ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() ) a__ : Optional[int] =["HuggingFace is in ", "Paris is in France"] a__ : Optional[Any] =text_classifier(lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}, {"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] , ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["label"] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format a__ : Union[str, Any] =text_classifier(lowerCAmelCase__ , top_k=lowerCAmelCase__ ) a__ : Optional[Any] =len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [[{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] * N, [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] * N] , ) a__ : List[str] ={"text": "HuggingFace is in ", "text_pair": "Paris is in France"} a__ : Optional[Any] =text_classifier(lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , {"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )} , ) self.assertTrue(outputs["label"] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. a__ : Any =[["HuggingFace is in ", "Paris is in France"]] with self.assertRaises(lowerCAmelCase__ ): text_classifier(lowerCAmelCase__ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility a__ : Optional[int] =text_classifier([[["HuggingFace is in ", "Paris is in France"]]] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] , ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() )
148
1
UpperCAmelCase_ : Dict = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCAmelCase_ : str = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def SCREAMING_SNAKE_CASE_ ( __magic_name__ : dict[int, list[int]] , __magic_name__ : int , __magic_name__ : list[bool] ) -> list[int]: """simple docstring""" UpperCamelCase :Any = True UpperCamelCase :int = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__magic_name__ , __magic_name__ , __magic_name__ ) order.append(__magic_name__ ) return order def SCREAMING_SNAKE_CASE_ ( __magic_name__ : dict[int, list[int]] , __magic_name__ : int , __magic_name__ : list[bool] ) -> list[int]: """simple docstring""" UpperCamelCase :Dict = True UpperCamelCase :Tuple = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__magic_name__ , __magic_name__ , __magic_name__ ) return component def SCREAMING_SNAKE_CASE_ ( __magic_name__ : dict[int, list[int]] ) -> list[list[int]]: """simple docstring""" UpperCamelCase :List[Any] = len(__magic_name__ ) * [False] UpperCamelCase :dict[int, list[int]] = {vert: [] for vert in range(len(__magic_name__ ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__magic_name__ ) UpperCamelCase :List[str] = [] for i, was_visited in enumerate(__magic_name__ ): if not was_visited: order += topology_sort(__magic_name__ , __magic_name__ , __magic_name__ ) UpperCamelCase :Optional[int] = [] UpperCamelCase :Dict = len(__magic_name__ ) * [False] for i in range(len(__magic_name__ ) ): UpperCamelCase :int = order[len(__magic_name__ ) - i - 1] if not visited[vert]: UpperCamelCase :Optional[int] = find_components(__magic_name__ , __magic_name__ , __magic_name__ ) components_list.append(__magic_name__ ) return components_list
38
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : Optional[Any] = TransfoXLTokenizer snake_case__ : List[Any] = False snake_case__ : Tuple = False def _A ( self : str ): super().setUp() UpperCamelCase :Dict = [ """<unk>""", """[CLS]""", """[SEP]""", """want""", """unwanted""", """wa""", """un""", """running""", """,""", """low""", """l""", ] UpperCamelCase :str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _A ( self : List[str] , **__lowerCamelCase : Any ): UpperCamelCase :Any = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _A ( self : Any , __lowerCamelCase : int ): UpperCamelCase :List[Any] = """<unk> UNwanted , running""" UpperCamelCase :int = """<unk> unwanted, running""" return input_text, output_text def _A ( self : Tuple ): UpperCamelCase :List[str] = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=__lowerCamelCase ) UpperCamelCase :Any = tokenizer.tokenize("""<unk> UNwanted , running""" ) self.assertListEqual(__lowerCamelCase , ["""<unk>""", """unwanted""", """,""", """running"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [0, 4, 8, 7] ) def _A ( self : Optional[Any] ): UpperCamelCase :List[Any] = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) def _A ( self : Union[str, Any] ): UpperCamelCase :int = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _A ( self : Tuple ): UpperCamelCase :Any = TransfoXLTokenizer(lower_case=__lowerCamelCase ) UpperCamelCase :Optional[int] = """Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?""" UpperCamelCase :Optional[int] = [ """Hello""", """(""", """bracket""", """)""", """and""", """side""", """@-@""", """scrolled""", """[""", """and""", """]""", """Henry""", """'s""", """$""", """5""", """@,@""", """000""", """with""", """3""", """@.@""", """34""", """m""", """.""", """What""", """'s""", """up""", """!""", """?""", ] self.assertListEqual(tokenizer.tokenize(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowerCamelCase ) , __lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Any = self.get_tokenizer() UpperCamelCase :List[str] = len(__lowerCamelCase ) tokenizer.add_tokens(["""new1""", """new2"""] ) tokenizer.move_added_token("""new1""" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowerCamelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("""new1""" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , """new1""" )
38
1
'''simple docstring''' import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a : List[str] = get_logger(__name__) class UpperCamelCase_ ( enum.Enum ): lowercase = 'all_checks' lowercase = 'basic_checks' lowercase = 'no_checks' class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass def __lowerCamelCase ( _lowercase , _lowercase , _lowercase=None ) -> Union[str, Any]: if expected_checksums is None: logger.info("""Unable to verify checksums.""" ) return if len(set(_lowercase ) - set(_lowercase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(_lowercase ) - set(_lowercase ) ) ) if len(set(_lowercase ) - set(_lowercase ) ) > 0: raise UnexpectedDownloadedFile(str(set(_lowercase ) - set(_lowercase ) ) ) UpperCAmelCase : str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] UpperCAmelCase : Dict = """ for """ + verification_name if verification_name is not None else """""" if len(_lowercase ) > 0: raise NonMatchingChecksumError( F'''Checksums didn\'t match{for_verification_name}:\n''' F'''{bad_urls}\n''' """Set `verification_mode='no_checks'` to skip checksums verification and ignore this error""" ) logger.info("""All the checksums matched successfully""" + for_verification_name ) class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass class UpperCamelCase_ ( __magic_name__ ): pass def __lowerCamelCase ( _lowercase , _lowercase ) -> Optional[Any]: if expected_splits is None: logger.info("""Unable to verify splits sizes.""" ) return if len(set(_lowercase ) - set(_lowercase ) ) > 0: raise ExpectedMoreSplits(str(set(_lowercase ) - set(_lowercase ) ) ) if len(set(_lowercase ) - set(_lowercase ) ) > 0: raise UnexpectedSplits(str(set(_lowercase ) - set(_lowercase ) ) ) UpperCAmelCase : Dict = [ {"""expected""": expected_splits[name], """recorded""": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_lowercase ) > 0: raise NonMatchingSplitsSizesError(str(_lowercase ) ) logger.info("""All the splits matched successfully.""" ) def __lowerCamelCase ( _lowercase , _lowercase = True ) -> dict: if record_checksum: UpperCAmelCase : Any = shaaaa() with open(_lowercase , """rb""" ) as f: for chunk in iter(lambda: f.read(1 << 2_0 ) , B"""""" ): m.update(_lowercase ) UpperCAmelCase : Union[str, Any] = m.hexdigest() else: UpperCAmelCase : Dict = None return {"num_bytes": os.path.getsize(_lowercase ), "checksum": checksum} def __lowerCamelCase ( _lowercase ) -> Union[str, Any]: if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
338
'''simple docstring''' from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class UpperCamelCase_ : lowercase = MBartConfig lowercase = {} lowercase = 'gelu' def __init__( self , A , A=13 , A=7 , A=True , A=False , A=99 , A=32 , A=2 , A=4 , A=37 , A=0.1 , A=0.1 , A=20 , A=2 , A=1 , A=0 , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Dict = batch_size UpperCAmelCase : Tuple = seq_length UpperCAmelCase : str = is_training UpperCAmelCase : Optional[int] = use_labels UpperCAmelCase : Optional[Any] = vocab_size UpperCAmelCase : Union[str, Any] = hidden_size UpperCAmelCase : Union[str, Any] = num_hidden_layers UpperCAmelCase : List[Any] = num_attention_heads UpperCAmelCase : Optional[int] = intermediate_size UpperCAmelCase : Dict = hidden_dropout_prob UpperCAmelCase : int = attention_probs_dropout_prob UpperCAmelCase : Optional[int] = max_position_embeddings UpperCAmelCase : Optional[Any] = eos_token_id UpperCAmelCase : List[str] = pad_token_id UpperCAmelCase : List[Any] = bos_token_id def _lowercase( self ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase : List[str] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : str = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase : List[Any] = prepare_mbart_inputs_dict(A , A , A ) return config, inputs_dict def _lowercase( self , A , A ) -> List[str]: UpperCAmelCase : List[str] = TFMBartModel(config=A ).get_decoder() UpperCAmelCase : int = inputs_dict["""input_ids"""] UpperCAmelCase : str = input_ids[:1, :] UpperCAmelCase : Optional[Any] = inputs_dict["""attention_mask"""][:1, :] UpperCAmelCase : List[str] = inputs_dict["""head_mask"""] UpperCAmelCase : List[Any] = 1 # first forward pass UpperCAmelCase : List[str] = model(A , attention_mask=A , head_mask=A , use_cache=A ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = outputs.to_tuple() UpperCAmelCase : int = past_key_values[1] def __lowerCamelCase ( _lowercase , _lowercase , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> List[str]: if attention_mask is None: UpperCAmelCase : Tuple = tf.cast(tf.math.not_equal(_lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase : int = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase : List[Any] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class UpperCamelCase_ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowercase = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowercase = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowercase = True lowercase = False lowercase = False def _lowercase( self , A , A , A , A , A ) -> int: if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def _lowercase( self ) -> Optional[Any]: UpperCAmelCase : int = TFMBartModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=A ) def _lowercase( self ) -> Optional[int]: self.config_tester.run_common_tests() def _lowercase( self ) -> Dict: UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A ) @require_sentencepiece @require_tokenizers @require_tf class UpperCamelCase_ ( unittest.TestCase ): lowercase = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowercase = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowercase = 'facebook/mbart-large-en-ro' @cached_property def _lowercase( self ) -> Any: return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def _lowercase( self ) -> List[Any]: UpperCAmelCase : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def _lowercase( self , **A ) -> Any: UpperCAmelCase : Optional[int] = self.translate_src_text(**A ) self.assertListEqual(self.expected_text , A ) def _lowercase( self , **A ) -> Optional[Any]: UpperCAmelCase : List[str] = self.tokenizer(self.src_text , **A , return_tensors="""tf""" ) UpperCAmelCase : int = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) UpperCAmelCase : Any = self.tokenizer.batch_decode(A , skip_special_tokens=A ) return generated_words @slow def _lowercase( self ) -> List[Any]: self._assert_generated_batch_equal_expected()
338
1
import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __A: @staticmethod def SCREAMING_SNAKE_CASE_ ( *_snake_case , **_snake_case ) -> Tuple: '''simple docstring''' pass @is_pipeline_test @require_vision @require_timm @require_torch class __A( unittest.TestCase ): snake_case_ = MODEL_FOR_OBJECT_DETECTION_MAPPING def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case ) -> List[str]: '''simple docstring''' __a = ObjectDetectionPipeline(model=_snake_case , image_processor=_snake_case ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[int]: '''simple docstring''' __a = object_detector('''./tests/fixtures/tests_samples/COCO/000000039769.png''' , threshold=0.0 ) self.assertGreater(len(_snake_case ) , 0 ) for detected_object in outputs: self.assertEqual( _snake_case , { '''score''': ANY(_snake_case ), '''label''': ANY(_snake_case ), '''box''': {'''xmin''': ANY(_snake_case ), '''ymin''': ANY(_snake_case ), '''xmax''': ANY(_snake_case ), '''ymax''': ANY(_snake_case )}, } , ) import datasets __a = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) __a = [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] __a = object_detector(_snake_case , threshold=0.0 ) self.assertEqual(len(_snake_case ) , len(_snake_case ) ) for outputs in batch_outputs: self.assertGreater(len(_snake_case ) , 0 ) for detected_object in outputs: self.assertEqual( _snake_case , { '''score''': ANY(_snake_case ), '''label''': ANY(_snake_case ), '''box''': {'''xmin''': ANY(_snake_case ), '''ymin''': ANY(_snake_case ), '''xmax''': ANY(_snake_case ), '''ymax''': ANY(_snake_case )}, } , ) @require_tf @unittest.skip('''Object detection not implemented in TF''' ) def SCREAMING_SNAKE_CASE_ ( self ) -> Dict: '''simple docstring''' pass @require_torch def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' __a = '''hf-internal-testing/tiny-detr-mobilenetsv3''' __a = AutoModelForObjectDetection.from_pretrained(_snake_case ) __a = AutoFeatureExtractor.from_pretrained(_snake_case ) __a = ObjectDetectionPipeline(model=_snake_case , feature_extractor=_snake_case ) __a = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=0.0 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ] , ) __a = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], ] , ) @require_torch @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Dict: '''simple docstring''' __a = '''facebook/detr-resnet-50''' __a = AutoModelForObjectDetection.from_pretrained(_snake_case ) __a = AutoFeatureExtractor.from_pretrained(_snake_case ) __a = ObjectDetectionPipeline(model=_snake_case , feature_extractor=_snake_case ) __a = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) __a = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ] , ) @require_torch @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' __a = '''facebook/detr-resnet-50''' __a = pipeline('''object-detection''' , model=_snake_case ) __a = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) __a = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ] , ) @require_torch @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' __a = 0.9985 __a = '''facebook/detr-resnet-50''' __a = pipeline('''object-detection''' , model=_snake_case ) __a = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=_snake_case ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) @require_torch @require_pytesseract @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' __a = '''Narsil/layoutlmv3-finetuned-funsd''' __a = 0.9993 __a = pipeline('''object-detection''' , model=_snake_case , threshold=_snake_case ) __a = object_detector( '''https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png''' ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [ {'''score''': 0.9993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, {'''score''': 0.9993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, ] , )
6
# flake8: noqa # Lint as: python3 A : Optional[Any] = [ 'VerificationMode', 'Version', 'disable_progress_bar', 'enable_progress_bar', 'is_progress_bar_enabled', 'experimental', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
6
1
'''simple docstring''' import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , *lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , **lowerCAmelCase_ ): """simple docstring""" super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) _snake_case = eval_examples _snake_case = post_process_function def lowerCamelCase ( self , lowerCAmelCase_ = None , lowerCAmelCase_=None , lowerCAmelCase_ = None , lowerCAmelCase_ = "eval" , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = gen_kwargs.copy() _snake_case = ( gen_kwargs['max_length'] if gen_kwargs.get('max_length' ) is not None else self.args.generation_max_length ) _snake_case = ( gen_kwargs['num_beams'] if gen_kwargs.get('num_beams' ) is not None else self.args.generation_num_beams ) _snake_case = gen_kwargs _snake_case = self.eval_dataset if eval_dataset is None else eval_dataset _snake_case = self.get_eval_dataloader(lowerCAmelCase_ ) _snake_case = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _snake_case = self.compute_metrics _snake_case = None _snake_case = time.time() _snake_case = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _snake_case = eval_loop( lowerCAmelCase_ , description='Evaluation' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCAmelCase_ , metric_key_prefix=lowerCAmelCase_ , ) finally: _snake_case = compute_metrics _snake_case = self.args.eval_batch_size * self.args.world_size if F'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[F'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( lowerCAmelCase_ , lowerCAmelCase_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default _snake_case = self.post_process_function(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.compute_metrics(lowerCAmelCase_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'{metric_key_prefix}_' ): _snake_case = metrics.pop(lowerCAmelCase_ ) metrics.update(output.metrics ) else: _snake_case = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowerCAmelCase_ ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) _snake_case = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowerCAmelCase_ ) return metrics def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_ = "test" , **lowerCAmelCase_ ): """simple docstring""" _snake_case = gen_kwargs.copy() _snake_case = self.get_test_dataloader(lowerCAmelCase_ ) # Temporarily disable metric computation, we will do it in the loop here. _snake_case = self.compute_metrics _snake_case = None _snake_case = time.time() _snake_case = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _snake_case = eval_loop( lowerCAmelCase_ , description='Prediction' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCAmelCase_ , metric_key_prefix=lowerCAmelCase_ , ) finally: _snake_case = compute_metrics _snake_case = self.args.eval_batch_size * self.args.world_size if F'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[F'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( lowerCAmelCase_ , lowerCAmelCase_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output _snake_case = self.post_process_function(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , 'predict' ) _snake_case = self.compute_metrics(lowerCAmelCase_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'{metric_key_prefix}_' ): _snake_case = metrics.pop(lowerCAmelCase_ ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowerCAmelCase_ )
160
'''simple docstring''' import torch from transformers import AutoModel class __UpperCAmelCase ( torch.nn.Module ): def __init__( self , lowerCAmelCase_="sayef/fsner-bert-base-uncased" ): """simple docstring""" super(lowerCAmelCase_ , self ).__init__() _snake_case = AutoModel.from_pretrained(lowerCAmelCase_ , return_dict=lowerCAmelCase_ ) _snake_case = torch.nn.CosineSimilarity(3 , 1E-08 ) _snake_case = torch.nn.Softmax(dim=1 ) def lowerCamelCase ( self , **lowerCAmelCase_ ): """simple docstring""" return self.bert(**lowerCAmelCase_ ).last_hidden_state def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return token_embeddings.sum(2 , keepdim=lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 ): """simple docstring""" return self.softmax(T * self.cos(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = W_supports['sizes'].tolist() _snake_case = W_supports['start_token_id'].item() _snake_case = W_supports['end_token_id'].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] _snake_case = self.BERT(**lowerCAmelCase_ ) _snake_case = self.BERT(**lowerCAmelCase_ ) _snake_case = None _snake_case = None _snake_case = W_supports['input_ids'] == start_token_id _snake_case = W_supports['input_ids'] == end_token_id for i, size in enumerate(lowerCAmelCase_ ): if i == 0: _snake_case = 0 else: _snake_case = support_sizes[i - 1] _snake_case = S[s : s + size][start_token_masks[s : s + size]] _snake_case = S[s : s + size][end_token_masks[s : s + size]] _snake_case = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) _snake_case = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: _snake_case = torch.vstack((p_starts, p_start) ) _snake_case = torch.vstack((p_ends, p_end) ) else: _snake_case = p_start _snake_case = p_end return p_starts, p_ends
160
1
from __future__ import annotations from random import random class A_ : def __init__( self : Any ,SCREAMING_SNAKE_CASE__ : int | None = None): __lowerCamelCase : Union[str, Any] = value __lowerCamelCase : int = random() __lowerCamelCase : Node | None = None __lowerCamelCase : Node | None = None def __repr__( self : List[Any]): from pprint import pformat if self.left is None and self.right is None: return F"'{self.value}: {self.prior:.5}'" else: return pformat( {F"{self.value}: {self.prior:.5}": (self.left, self.right)} ,indent=1) def __str__( self : Union[str, Any]): __lowerCamelCase : Optional[int] = str(self.value) + ' ' __lowerCamelCase : Optional[Any] = str(self.left or '') __lowerCamelCase : List[Any] = str(self.right or '') return value + left + right def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> tuple[Node | None, Node | None]: if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: __lowerCamelCase , __lowerCamelCase : Union[str, Any] = split(root.left , lowerCamelCase__ ) return left, root else: __lowerCamelCase , __lowerCamelCase : List[str] = split(root.right , lowerCamelCase__ ) return root, right def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Node | None: if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: __lowerCamelCase : Optional[Any] = merge(left.right , lowerCamelCase__ ) return left else: __lowerCamelCase : List[str] = merge(lowerCamelCase__ , right.left ) return right def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Node | None: __lowerCamelCase : List[Any] = Node(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Optional[int] = split(lowerCamelCase__ , lowerCamelCase__ ) return merge(merge(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Node | None: __lowerCamelCase , __lowerCamelCase : List[str] = split(lowerCamelCase__ , value - 1 ) __lowerCamelCase , __lowerCamelCase : Any = split(lowerCamelCase__ , lowerCamelCase__ ) return merge(lowerCamelCase__ , lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> None: if not root: # None return else: inorder(root.left ) print(root.value , end=',' ) inorder(root.right ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Node | None: for arg in args.split(): if arg[0] == "+": __lowerCamelCase : str = insert(lowerCamelCase__ , int(arg[1:] ) ) elif arg[0] == "-": __lowerCamelCase : List[Any] = erase(lowerCamelCase__ , int(arg[1:] ) ) else: print('Unknown command' ) return root def SCREAMING_SNAKE_CASE__ ( ) -> None: __lowerCamelCase : int = None print( 'enter numbers to create a tree, + value to add value into treap, ' '- value to erase all nodes with value. \'q\' to quit. ' ) __lowerCamelCase : Optional[int] = input() while args != "q": __lowerCamelCase : Optional[int] = interact_treap(lowerCamelCase__ , lowerCamelCase__ ) print(lowerCamelCase__ ) __lowerCamelCase : int = input() print('good by!' ) if __name__ == "__main__": import doctest doctest.testmod() main()
73
import os from datetime import datetime as dt from github import Github A_ :str = [ '''good first issue''', '''feature request''', '''wip''', ] def A ( ) -> Any: __UpperCamelCase : Any =Github(os.environ['GITHUB_TOKEN'] ) __UpperCamelCase : Union[str, Any] =g.get_repo('huggingface/accelerate' ) __UpperCamelCase : Tuple =repo.get_issues(state='open' ) for issue in open_issues: __UpperCamelCase : List[Any] =sorted([comment for comment in issue.get_comments()] ,key=lambda a_ : i.created_at ,reverse=a_ ) __UpperCamelCase : str =comments[0] if len(a_ ) > 0 else None __UpperCamelCase : Any =dt.utcnow() __UpperCamelCase : List[str] =(current_time - issue.updated_at).days __UpperCamelCase : Union[str, Any] =(current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='closed' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
71
0
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger __lowerCamelCase = """<<<<<<< This should probably be modified because it mentions: """ __lowerCamelCase = """======= >>>>>>> """ __lowerCamelCase = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] __lowerCamelCase = [ # (pattern, replacement) # Order is important here for some replacements (R"""tfds\.core""", R"""datasets"""), (R"""tf\.io\.gfile\.GFile""", R"""open"""), (R"""tf\.([\w\d]+)""", R"""datasets.Value('\1')"""), (R"""tfds\.features\.Text\(\)""", R"""datasets.Value('string')"""), (R"""tfds\.features\.Text\(""", R"""datasets.Value('string'),"""), (R"""features\s*=\s*tfds.features.FeaturesDict\(""", R"""features=datasets.Features("""), (R"""tfds\.features\.FeaturesDict\(""", R"""dict("""), (R"""The TensorFlow Datasets Authors""", R"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (R"""tfds\.""", R"""datasets."""), (R"""dl_manager\.manual_dir""", R"""self.config.data_dir"""), (R"""self\.builder_config""", R"""self.config"""), ] def UpperCamelCase ( __lowerCamelCase : Namespace ): return ConvertCommand(args.tfds_path , args.datasets_directory ) class UpperCAmelCase ( A_ ): @staticmethod def _SCREAMING_SNAKE_CASE (snake_case__ : ArgumentParser ) -> Union[str, Any]: '''simple docstring''' snake_case : int = parser.add_parser( "convert" , help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset." , ) train_parser.add_argument( "--tfds_path" , type=snake_case__ , required=snake_case__ , help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert." , ) train_parser.add_argument( "--datasets_directory" , type=snake_case__ , required=snake_case__ , help="Path to the HuggingFace Datasets folder." ) train_parser.set_defaults(func=snake_case__ ) def __init__(self : str , snake_case__ : str , snake_case__ : str , *snake_case__ : Any ) -> Tuple: '''simple docstring''' snake_case : Union[str, Any] = get_logger("datasets-cli/converting" ) snake_case : int = tfds_path snake_case : Optional[Any] = datasets_directory def _SCREAMING_SNAKE_CASE (self : str ) -> Tuple: '''simple docstring''' if os.path.isdir(self._tfds_path ): snake_case : Union[str, Any] = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): snake_case : str = os.path.dirname(self._tfds_path ) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path." ) snake_case : int = os.path.abspath(self._datasets_directory ) self._logger.info(f"""Converting datasets from {abs_tfds_path} to {abs_datasets_path}""" ) snake_case : Optional[int] = [] snake_case : Any = [] snake_case : Optional[int] = {} if os.path.isdir(self._tfds_path ): snake_case : Tuple = os.listdir(snake_case__ ) else: snake_case : int = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f"""Looking at file {f_name}""" ) snake_case : List[str] = os.path.join(snake_case__ , snake_case__ ) snake_case : List[str] = os.path.join(snake_case__ , snake_case__ ) if not os.path.isfile(snake_case__ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file" ) continue with open(snake_case__ , encoding="utf-8" ) as f: snake_case : List[str] = f.readlines() snake_case : List[Any] = [] snake_case : Optional[int] = False snake_case : Tuple = False snake_case : Union[str, Any] = [] for line in lines: snake_case : List[str] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: snake_case : Union[str, Any] = "import datasets\n" elif "import tensorflow" in out_line: # order is important here snake_case : Optional[Any] = "" continue elif "from absl import logging" in out_line: snake_case : Optional[int] = "from datasets import logging\n" elif "getLogger" in out_line: snake_case : Optional[Any] = out_line.replace("getLogger" , "get_logger" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): snake_case : List[str] = True snake_case : Dict = list(filter(lambda snake_case__ : e in out_line , snake_case__ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(snake_case__ ) + "\n" ) out_lines.append(snake_case__ ) out_lines.append(snake_case__ ) continue else: for pattern, replacement in TO_CONVERT: snake_case : List[Any] = re.sub(snake_case__ , snake_case__ , snake_case__ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: snake_case : List[Any] = re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)" , snake_case__ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split("," ) ) snake_case : Any = "from . import " + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f"""Error converting {out_line.strip()}""" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: snake_case : List[str] = True out_lines.append(snake_case__ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset snake_case : Dict = f_name.replace(".py" , "" ) snake_case : str = os.path.join(snake_case__ , snake_case__ ) snake_case : Union[str, Any] = os.path.join(snake_case__ , snake_case__ ) os.makedirs(snake_case__ , exist_ok=snake_case__ ) self._logger.info(f"""Adding directory {output_dir}""" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(snake_case__ ) if needs_manual_update: with_manual_update.append(snake_case__ ) with open(snake_case__ , "w" , encoding="utf-8" ) as f: f.writelines(snake_case__ ) self._logger.info(f"""Converted in {output_file}""" ) for utils_file in utils_files: try: snake_case : List[str] = os.path.basename(snake_case__ ) snake_case : Any = imports_to_builder_map[f_name.replace(".py" , "" )] self._logger.info(f"""Moving {dest_folder} to {utils_file}""" ) shutil.copy(snake_case__ , snake_case__ ) except KeyError: self._logger.error(f"""Cannot find destination folder for {utils_file}. Please copy manually.""" ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f"""You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'.""" )
10
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor __lowerCamelCase = logging.get_logger(__name__) class UpperCAmelCase ( A_ ): def __init__(self : List[Any] , *snake_case__ : List[str] , **snake_case__ : Dict ) -> None: '''simple docstring''' warnings.warn( "The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use PerceiverImageProcessor instead." , snake_case__ , ) super().__init__(*snake_case__ , **snake_case__ )
10
1
"""simple docstring""" import baseaa def lowercase ( _snake_case : str ) ->bytes: """simple docstring""" return baseaa.baaencode(string.encode('''utf-8''' ) ) def lowercase ( _snake_case : bytes ) ->str: """simple docstring""" return baseaa.baadecode(_snake_case ).decode('''utf-8''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[Any] = """Hello World!""" SCREAMING_SNAKE_CASE : Optional[Any] = baseaa_encode(test) print(encoded) SCREAMING_SNAKE_CASE : str = baseaa_decode(encoded) print(decoded)
102
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=UpperCamelCase) class snake_case__ ( UpperCamelCase): a_ = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) a_ = Features({"text": Value("string")}) a_ = Features({}) a_ = "text" @property def A ( self : List[str] ) -> Dict[str, str]: return {self.text_column: "text"}
304
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : str = logging.get_logger(__name__) _lowerCAmelCase : List[Any] = { "google/pix2struct-textcaps-base": ( "https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json" ), } class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : int = 'pix2struct_text_model' __SCREAMING_SNAKE_CASE : str = ['past_key_values'] __SCREAMING_SNAKE_CASE : int = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : Optional[Any] , A : Tuple=5_0_2_4_4 , A : Optional[Any]=7_6_8 , A : str=6_4 , A : int=2_0_4_8 , A : Tuple=1_2 , A : Union[str, Any]=1_2 , A : List[Any]=3_2 , A : Dict=1_2_8 , A : Optional[Any]=0.1 , A : int=1e-6 , A : int=1.0 , A : Dict="gelu_new" , A : List[str]=0 , A : Optional[int]=False , A : Optional[Any]=0 , A : Union[str, Any]=1 , A : str=False , A : List[Any]=True , **A : Optional[Any] , ): _UpperCAmelCase : str = vocab_size _UpperCAmelCase : Union[str, Any] = hidden_size _UpperCAmelCase : Tuple = d_kv _UpperCAmelCase : Optional[Any] = d_ff _UpperCAmelCase : Union[str, Any] = num_layers _UpperCAmelCase : Union[str, Any] = num_heads _UpperCAmelCase : Tuple = relative_attention_num_buckets _UpperCAmelCase : str = relative_attention_max_distance _UpperCAmelCase : Union[str, Any] = dropout_rate _UpperCAmelCase : str = layer_norm_epsilon _UpperCAmelCase : Union[str, Any] = initializer_factor _UpperCAmelCase : Dict = use_cache _UpperCAmelCase : List[str] = eos_token_id _UpperCAmelCase : Optional[Any] = decoder_start_token_id # for backwards compatibility _UpperCAmelCase : Optional[Any] = dense_act_fn super().__init__( pad_token_id=A , eos_token_id=A , decoder_start_token_id=A , tie_word_embeddings=A , is_decoder=A , **A , ) @classmethod def snake_case_ ( cls : Optional[Any] , A : Union[str, os.PathLike] , **A : Tuple ): cls._set_token_in_kwargs(A ) _UpperCAmelCase , _UpperCAmelCase : Dict = cls.get_config_dict(A , **A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("model_type" ) == "pix2struct": _UpperCAmelCase : List[str] = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A , **A ) class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : str = 'pix2struct_vision_model' def __init__( self : Tuple , A : Optional[Any]=7_6_8 , A : str=7_6_8 , A : Any=2_0_4_8 , A : Optional[int]=6_4 , A : Tuple=1_2 , A : Tuple=1_2 , A : List[str]="gelu_new" , A : Any=1e-6 , A : Union[str, Any]=0.0 , A : Dict=0.0 , A : Optional[Any]=1e-10 , A : List[str]=1.0 , A : Any=4_0_9_6 , A : Optional[int]=3_2 , A : Optional[int]=1_2_8 , **A : Optional[int] , ): super().__init__(**A ) _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : str = patch_embed_hidden_size _UpperCAmelCase : Tuple = d_ff _UpperCAmelCase : List[Any] = dropout_rate _UpperCAmelCase : Tuple = num_hidden_layers _UpperCAmelCase : str = num_attention_heads _UpperCAmelCase : str = initializer_range _UpperCAmelCase : Optional[int] = initializer_factor _UpperCAmelCase : Union[str, Any] = attention_dropout _UpperCAmelCase : Optional[Any] = layer_norm_eps _UpperCAmelCase : Tuple = dense_act_fn _UpperCAmelCase : Tuple = seq_len _UpperCAmelCase : Union[str, Any] = relative_attention_num_buckets _UpperCAmelCase : Any = relative_attention_max_distance _UpperCAmelCase : Optional[Any] = d_kv @classmethod def snake_case_ ( cls : Optional[Any] , A : Union[str, os.PathLike] , **A : Any ): cls._set_token_in_kwargs(A ) _UpperCAmelCase , _UpperCAmelCase : Optional[int] = cls.get_config_dict(A , **A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("model_type" ) == "pix2struct": _UpperCAmelCase : int = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A , **A ) class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : Optional[int] = 'pix2struct' __SCREAMING_SNAKE_CASE : Dict = True def __init__( self : Any , A : List[Any]=None , A : Optional[int]=None , A : Tuple=1.0 , A : Union[str, Any]=0.02 , A : Optional[int]=False , A : int=False , A : List[Any]=True , **A : Optional[int] , ): super().__init__(tie_word_embeddings=A , is_encoder_decoder=A , **A ) if text_config is None: _UpperCAmelCase : int = {} logger.info("text_config is None. Initializing the Pix2StructTextConfig with default values." ) if vision_config is None: _UpperCAmelCase : List[str] = {} logger.info("vision_config is None. Initializing the Pix2StructVisionConfig with default values." ) _UpperCAmelCase : Tuple = PixaStructTextConfig(**A ) _UpperCAmelCase : Optional[Any] = PixaStructVisionConfig(**A ) _UpperCAmelCase : Any = self.text_config.decoder_start_token_id _UpperCAmelCase : Optional[Any] = self.text_config.pad_token_id _UpperCAmelCase : Any = self.text_config.eos_token_id _UpperCAmelCase : int = initializer_factor _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Tuple = self.initializer_range _UpperCAmelCase : Union[str, Any] = self.initializer_range _UpperCAmelCase : int = is_vqa @classmethod def snake_case_ ( cls : Optional[Any] , A : PixaStructTextConfig , A : PixaStructVisionConfig , **A : Union[str, Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **A ) def snake_case_ ( self : Tuple ): _UpperCAmelCase : Optional[int] = copy.deepcopy(self.__dict__ ) _UpperCAmelCase : Dict = self.text_config.to_dict() _UpperCAmelCase : Any = self.vision_config.to_dict() _UpperCAmelCase : Optional[int] = self.__class__.model_type return output
202
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _lowerCAmelCase : Tuple = logging.get_logger(__name__) class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : str = ['pixel_values'] def __init__( self : List[Any] , A : bool = True , A : Dict[str, int] = None , A : PILImageResampling = PILImageResampling.BICUBIC , A : bool = True , A : Dict[str, int] = None , A : bool = True , A : Union[int, float] = 1 / 2_5_5 , A : bool = True , A : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_MEAN , A : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_STD , **A : List[Any] , ): super().__init__(**A ) _UpperCAmelCase : int = size if size is not None else {"shortest_edge": 2_2_4} _UpperCAmelCase : Tuple = get_size_dict(A , default_to_square=A ) _UpperCAmelCase : int = crop_size if crop_size is not None else {"height": 2_2_4, "width": 2_2_4} _UpperCAmelCase : Union[str, Any] = get_size_dict(A , param_name="crop_size" ) _UpperCAmelCase : str = do_resize _UpperCAmelCase : Union[str, Any] = size _UpperCAmelCase : Tuple = resample _UpperCAmelCase : List[Any] = do_center_crop _UpperCAmelCase : List[Any] = crop_size _UpperCAmelCase : Tuple = do_rescale _UpperCAmelCase : List[Any] = rescale_factor _UpperCAmelCase : Dict = do_normalize _UpperCAmelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN _UpperCAmelCase : Any = image_std if image_std is not None else IMAGENET_DEFAULT_STD def snake_case_ ( self : Optional[int] , A : np.ndarray , A : Dict[str, int] , A : PILImageResampling = PILImageResampling.BICUBIC , A : Optional[Union[str, ChannelDimension]] = None , **A : Union[str, Any] , ): _UpperCAmelCase : Tuple = get_size_dict(A , default_to_square=A ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: _UpperCAmelCase : Any = int((2_5_6 / 2_2_4) * size["shortest_edge"] ) _UpperCAmelCase : Union[str, Any] = get_resize_output_image_size(A , size=A , default_to_square=A ) _UpperCAmelCase : str = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f'Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}' ) return resize( A , size=(size_dict["height"], size_dict["width"]) , resample=A , data_format=A , **A ) def snake_case_ ( self : int , A : np.ndarray , A : Dict[str, int] , A : Optional[Union[str, ChannelDimension]] = None , **A : Optional[Any] , ): _UpperCAmelCase : Dict = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(f'Size dict must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(A , size=(size["height"], size["width"]) , data_format=A , **A ) def snake_case_ ( self : List[str] , A : np.ndarray , A : Union[int, float] , A : Optional[Union[str, ChannelDimension]] = None , **A : Dict , ): return rescale(A , scale=A , data_format=A , **A ) def snake_case_ ( self : List[str] , A : np.ndarray , A : Union[float, List[float]] , A : Union[float, List[float]] , A : Optional[Union[str, ChannelDimension]] = None , **A : str , ): return normalize(A , mean=A , std=A , data_format=A , **A ) def snake_case_ ( self : Tuple , A : ImageInput , A : Optional[bool] = None , A : Optional[Dict[str, int]] = None , A : PILImageResampling = None , A : Optional[bool] = None , A : Optional[Dict[str, int]] = None , A : Optional[bool] = None , A : Optional[float] = None , A : Optional[bool] = None , A : Optional[Union[float, Iterable[float]]] = None , A : Optional[Union[float, Iterable[float]]] = None , A : Optional[TensorType] = None , A : ChannelDimension = ChannelDimension.FIRST , **A : Dict , ): _UpperCAmelCase : Dict = do_resize if do_resize is not None else self.do_resize _UpperCAmelCase : List[str] = resample if resample is not None else self.resample _UpperCAmelCase : int = do_center_crop if do_center_crop is not None else self.do_center_crop _UpperCAmelCase : str = do_rescale if do_rescale is not None else self.do_rescale _UpperCAmelCase : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCAmelCase : Optional[int] = do_normalize if do_normalize is not None else self.do_normalize _UpperCAmelCase : List[str] = image_mean if image_mean is not None else self.image_mean _UpperCAmelCase : Tuple = image_std if image_std is not None else self.image_std _UpperCAmelCase : Any = size if size is not None else self.size _UpperCAmelCase : Optional[int] = get_size_dict(A , default_to_square=A ) _UpperCAmelCase : List[str] = crop_size if crop_size is not None else self.crop_size _UpperCAmelCase : List[str] = get_size_dict(A , param_name="crop_size" ) _UpperCAmelCase : List[Any] = make_list_of_images(A ) if not valid_images(A ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _UpperCAmelCase : Any = [to_numpy_array(A ) for image in images] if do_resize: _UpperCAmelCase : str = [self.resize(A , A , A ) for image in images] if do_center_crop: _UpperCAmelCase : Dict = [self.center_crop(A , A ) for image in images] if do_rescale: _UpperCAmelCase : Dict = [self.rescale(A , A ) for image in images] if do_normalize: _UpperCAmelCase : Optional[int] = [self.normalize(A , A , A ) for image in images] _UpperCAmelCase : Union[str, Any] = [to_channel_dimension_format(A , A ) for image in images] _UpperCAmelCase : List[Any] = {"pixel_values": images} return BatchFeature(data=A , tensor_type=A )
202
1
'''simple docstring''' import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm _A : Any =re.compile('''[^A-Za-z_0-9]''') # parameters used in DuplicationIndex _A : Dict =10 _A : Tuple =256 def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Optional[MinHash]: if len(snake_case__ ) < MIN_NUM_TOKENS: return None lowerCamelCase__ : Any = MinHash(num_perm=snake_case__ ) for token in set(snake_case__ ): min_hash.update(token.encode() ) return min_hash def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Set[str]: return {t for t in NON_ALPHA.split(snake_case__ ) if len(t.strip() ) > 0} class _lowercase : def __init__( self: Optional[Any] , *, UpperCamelCase__: float = 0.85 , ): lowerCamelCase__ : Dict = duplication_jaccard_threshold lowerCamelCase__ : Union[str, Any] = NUM_PERM lowerCamelCase__ : Tuple = MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm ) lowerCamelCase__ : int = defaultdict(UpperCAmelCase_ ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Tuple , UpperCamelCase__: MinHash ): lowerCamelCase__ : List[str] = self._index.query(UpperCAmelCase_ ) if code_key in self._index.keys: print(F'''Duplicate key {code_key}''' ) return self._index.insert(UpperCAmelCase_ , UpperCAmelCase_ ) if len(UpperCAmelCase_ ) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(UpperCAmelCase_ ) break else: self._duplicate_clusters[close_duplicates[0]].add(UpperCAmelCase_ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : List[str] = [] for base, duplicates in self._duplicate_clusters.items(): lowerCamelCase__ : List[Any] = [base] + list(UpperCAmelCase_ ) # reformat the cluster to be a list of dict lowerCamelCase__ : Any = [{"""base_index""": el[0], """repo_name""": el[1], """path""": el[2]} for el in cluster] duplicate_clusters.append(UpperCAmelCase_ ) return duplicate_clusters def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_duplicate_clusters() with open(UpperCAmelCase_ , """w""" ) as f: json.dump(UpperCAmelCase_ , UpperCAmelCase_ ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[str]: lowerCamelCase__ , lowerCamelCase__ : Dict = element lowerCamelCase__ : int = get_min_hash([t for t in NON_ALPHA.split(data["""content"""] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(snake_case__ , max_queue_size=10000 ) , chunksize=100 , ): if data is not None: yield data def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = DuplicationIndex(duplication_jaccard_threshold=snake_case__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(snake_case__ ) ) , max_queue_size=100 ) ): di.add(snake_case__ , snake_case__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : List[Any] = get_tokens(snake_case__ ) lowerCamelCase__ : Optional[int] = get_tokens(snake_case__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) _A : str =None def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> str: lowerCamelCase__ : List[Any] = [] for elementa in cluster: lowerCamelCase__ : List[str] = _shared_dataset[elementa["""base_index"""]]["""content"""] for elementa in extremes: lowerCamelCase__ : Optional[Any] = _shared_dataset[elementa["""base_index"""]]["""content"""] if jaccard_similarity(snake_case__ , snake_case__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: lowerCamelCase__ : Dict = 1 extremes.append(snake_case__ ) return extremes def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[Any]: global _shared_dataset lowerCamelCase__ : Optional[int] = dataset lowerCamelCase__ : str = [] lowerCamelCase__ : List[str] = partial(_find_cluster_extremes_shared , jaccard_threshold=snake_case__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( snake_case__ , snake_case__ , ) , total=len(snake_case__ ) , ): extremes_list.append(snake_case__ ) return extremes_list def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 0.85 ) -> Tuple[Type[Dataset], List[List[Dict]]]: lowerCamelCase__ : Tuple = make_duplicate_clusters(snake_case__ , snake_case__ ) lowerCamelCase__ : Any = {x["""base_index"""] for cluster in duplicate_clusters for x in cluster} lowerCamelCase__ : Optional[Any] = {} lowerCamelCase__ : List[str] = find_extremes(snake_case__ , snake_case__ , snake_case__ ) for extremes in extremes_clusters: for element in extremes: lowerCamelCase__ : Any = element lowerCamelCase__ : Optional[Any] = duplicate_indices - set(extreme_dict.keys() ) lowerCamelCase__ : int = dataset.filter(lambda UpperCamelCase , UpperCamelCase : idx not in remove_indices , with_indices=snake_case__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: lowerCamelCase__ : Tuple = element["""base_index"""] in extreme_dict if element["is_extreme"]: lowerCamelCase__ : Optional[Any] = extreme_dict[element["""base_index"""]]["""copies"""] print(f'''Original dataset size: {len(snake_case__ )}''' ) print(f'''Number of duplicate clusters: {len(snake_case__ )}''' ) print(f'''Files in duplicate cluster: {len(snake_case__ )}''' ) print(f'''Unique files in duplicate cluster: {len(snake_case__ )}''' ) print(f'''Filtered dataset size: {len(snake_case__ )}''' ) return ds_filter, duplicate_clusters
41
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [randint(-10_00 ,10_00 ) for i in range(10 )] _SCREAMING_SNAKE_CASE = randint(-50_00 ,50_00 ) return (arr, r) UpperCamelCase = make_dataset() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(snake_case__ ,3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(n - 1 ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ) -> tuple[float, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _SCREAMING_SNAKE_CASE = """ triplet_sum1(*dataset) """ _SCREAMING_SNAKE_CASE = """ triplet_sum2(*dataset) """ _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
306
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __a (UpperCamelCase_ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :str = KandinskyVaaControlnetPipeline _SCREAMING_SNAKE_CASE :Optional[int] = ["""image_embeds""", """negative_image_embeds""", """hint"""] _SCREAMING_SNAKE_CASE :int = ["""image_embeds""", """negative_image_embeds""", """hint"""] _SCREAMING_SNAKE_CASE :Dict = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] _SCREAMING_SNAKE_CASE :Tuple = False @property def _a ( self ) -> Any: """simple docstring""" return 32 @property def _a ( self ) -> Union[str, Any]: """simple docstring""" return 32 @property def _a ( self ) -> List[str]: """simple docstring""" return self.time_input_dim @property def _a ( self ) -> str: """simple docstring""" return self.time_input_dim * 4 @property def _a ( self ) -> Optional[Any]: """simple docstring""" return 100 @property def _a ( self ) -> Any: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Any = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } SCREAMING_SNAKE_CASE__ : Optional[int] = UNetaDConditionModel(**_a ) return model @property def _a ( self ) -> Optional[Any]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def _a ( self ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : List[str] = VQModel(**self.dummy_movq_kwargs ) return model def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.dummy_unet SCREAMING_SNAKE_CASE__ : Optional[Any] = self.dummy_movq SCREAMING_SNAKE_CASE__ : Any = DDIMScheduler( num_train_timesteps=1_000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=_a , set_alpha_to_one=_a , steps_offset=1 , prediction_type="""epsilon""" , thresholding=_a , ) SCREAMING_SNAKE_CASE__ : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def _a ( self , _a , _a=0 ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_a ) ).to(_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _a ) # create hint SCREAMING_SNAKE_CASE__ : Dict = floats_tensor((1, 3, 64, 64) , rng=random.Random(_a ) ).to(_a ) if str(_a ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.manual_seed(_a ) else: SCREAMING_SNAKE_CASE__ : Tuple = torch.Generator(device=_a ).manual_seed(_a ) SCREAMING_SNAKE_CASE__ : Tuple = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = """cpu""" SCREAMING_SNAKE_CASE__ : Any = self.get_dummy_components() SCREAMING_SNAKE_CASE__ : str = self.pipeline_class(**_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : Dict = pipe(**self.get_dummy_inputs(_a ) ) SCREAMING_SNAKE_CASE__ : Optional[int] = output.images SCREAMING_SNAKE_CASE__ : Any = pipe( **self.get_dummy_inputs(_a ) , return_dict=_a , )[0] SCREAMING_SNAKE_CASE__ : int = image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE__ : Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE__ : Tuple = np.array( [0.6_959_826, 0.868_279, 0.7_558_092, 0.68_769_467, 0.85_805_804, 0.65_977_496, 0.44_885_302, 0.5_959_111, 0.4_251_595] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class __a (unittest.TestCase): '''simple docstring''' def _a ( self ) -> Any: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) SCREAMING_SNAKE_CASE__ : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) SCREAMING_SNAKE_CASE__ : int = torch.from_numpy(np.array(_a ) ).float() / 255.0 SCREAMING_SNAKE_CASE__ : List[Any] = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE__ : Dict = pipeline.to(_a ) pipeline.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : Tuple = """A robot, 4k photo""" SCREAMING_SNAKE_CASE__ : Dict = torch.Generator(device="""cuda""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = pipe_prior( _a , generator=_a , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() SCREAMING_SNAKE_CASE__ : int = torch.Generator(device="""cuda""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = pipeline( image_embeds=_a , negative_image_embeds=_a , hint=_a , generator=_a , num_inference_steps=100 , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : List[Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(_a , _a )
56
"""simple docstring""" class __a : '''simple docstring''' def __init__( self , _a ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = arr.split(""",""" ) def _a ( self ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = [int(self.array[0] )] * len(self.array ) SCREAMING_SNAKE_CASE__ : List[str] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): SCREAMING_SNAKE_CASE__ : Dict = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) SCREAMING_SNAKE_CASE__ : List[Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": a :Optional[Any] = input("please input some numbers:") a :Optional[Any] = SubArray(whole_array) a :Optional[int] = array.solve_sub_array() print(("the results is:", re))
56
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a__ : Optional[int] =logging.get_logger(__name__) a__ : Optional[int] ={ "kssteven/ibert-roberta-base": "https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json", "kssteven/ibert-roberta-large": "https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json", "kssteven/ibert-roberta-large-mnli": ( "https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json" ), } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] ="ibert" def __init__( self : Dict , __A : Union[str, Any]=3_0_5_2_2 , __A : List[Any]=7_6_8 , __A : List[Any]=1_2 , __A : Optional[int]=1_2 , __A : Optional[Any]=3_0_7_2 , __A : int="gelu" , __A : str=0.1 , __A : List[Any]=0.1 , __A : Optional[Any]=5_1_2 , __A : int=2 , __A : Union[str, Any]=0.02 , __A : List[str]=1e-12 , __A : Optional[int]=1 , __A : Optional[int]=0 , __A : List[str]=2 , __A : str="absolute" , __A : Any=False , __A : Optional[Any]="none" , **__A : Any , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = quant_mode __UpperCamelCase = force_dequant class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : Tuple ): if self.task == "multiple-choice": __UpperCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __UpperCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
53
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __UpperCAmelCase : Union[str, Any] = logging.getLogger(__name__) class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = """token-classification""" def __init__( self : Union[str, Any] , A : List[Any] ): if type(A ) == dict: __snake_case: str = Namespace(**A ) __snake_case: str = import_module("""tasks""" ) try: __snake_case: Tuple = getattr(A , hparams.task_type ) __snake_case: TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) __snake_case: Optional[int] = self.token_classification_task.get_labels(hparams.labels ) __snake_case: str = CrossEntropyLoss().ignore_index super().__init__(A , len(self.labels ) , self.mode ) def UpperCAmelCase__ ( self : Union[str, Any] , **A : Union[str, Any] ): return self.model(**A ) def UpperCAmelCase__ ( self : Any , A : Optional[int] , A : str ): __snake_case: str = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": __snake_case: List[str] = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids __snake_case: Any = self(**A ) __snake_case: Any = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def UpperCAmelCase__ ( self : List[str] ): __snake_case: Tuple = self.hparams for mode in ["train", "dev", "test"]: __snake_case: Union[str, Any] = self._feature_file(A ) if os.path.exists(A ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , A ) __snake_case: Tuple = torch.load(A ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) __snake_case: List[Any] = self.token_classification_task.read_examples_from_file(args.data_dir , A ) __snake_case: Optional[int] = self.token_classification_task.convert_examples_to_features( A , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["""xlnet"""] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["""xlnet"""] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=A , pad_on_left=bool(self.config.model_type in ["""xlnet"""] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("""Saving features into cached file %s""" , A ) torch.save(A , A ) def UpperCAmelCase__ ( self : List[str] , A : int , A : int , A : bool = False ): __snake_case: List[str] = self._feature_file(A ) logger.info("""Loading features from cached file %s""" , A ) __snake_case: int = torch.load(A ) __snake_case: Optional[int] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) __snake_case: Optional[Any] = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: __snake_case: Dict = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: __snake_case: Tuple = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) __snake_case: Optional[int] = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(A , A , A , A ) , batch_size=A ) def UpperCAmelCase__ ( self : Tuple , A : Optional[Any] , A : int ): """Compute validation""" "" __snake_case: List[str] = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": __snake_case: Union[str, Any] = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids __snake_case: List[str] = self(**A ) __snake_case , __snake_case: int = outputs[:2] __snake_case: List[str] = logits.detach().cpu().numpy() __snake_case: Any = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCAmelCase__ ( self : Dict , A : Dict ): __snake_case: Union[str, Any] = torch.stack([x["""val_loss"""] for x in outputs] ).mean() __snake_case: Tuple = np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) __snake_case: List[str] = np.argmax(A , axis=2 ) __snake_case: Optional[Any] = np.concatenate([x["""target"""] for x in outputs] , axis=0 ) __snake_case: Tuple = dict(enumerate(self.labels ) ) __snake_case: Dict = [[] for _ in range(out_label_ids.shape[0] )] __snake_case: Tuple = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) __snake_case: Dict = { """val_loss""": val_loss_mean, """accuracy_score""": accuracy_score(A , A ), """precision""": precision_score(A , A ), """recall""": recall_score(A , A ), """f1""": fa_score(A , A ), } __snake_case: Dict = dict(results.items() ) __snake_case: Dict = results return ret, preds_list, out_label_list def UpperCAmelCase__ ( self : Tuple , A : int ): # when stable __snake_case , __snake_case , __snake_case: List[str] = self._eval_end(A ) __snake_case: int = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCAmelCase__ ( self : Dict , A : Optional[int] ): # updating to test_epoch_end instead of deprecated test_end __snake_case , __snake_case , __snake_case: int = self._eval_end(A ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 __snake_case: List[Any] = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCAmelCase__ ( A : Any , A : int ): # Add NER specific options BaseTransformer.add_model_specific_args(A , A ) parser.add_argument( """--task_type""" , default="""NER""" , type=A , help="""Task type to fine tune in training (e.g. NER, POS, etc)""" ) parser.add_argument( """--max_seq_length""" , default=128 , type=A , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--labels""" , default="""""" , type=A , help="""Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.""" , ) parser.add_argument( """--gpus""" , default=0 , type=A , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser if __name__ == "__main__": __UpperCAmelCase : Tuple = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __UpperCAmelCase : Tuple = NERTransformer.add_model_specific_args(parser, os.getcwd()) __UpperCAmelCase : List[str] = parser.parse_args() __UpperCAmelCase : Optional[int] = NERTransformer(args) __UpperCAmelCase : Tuple = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __UpperCAmelCase : int = sorted(glob.glob(os.path.join(args.output_dir, "checkpoint-epoch=*.ckpt"), recursive=True)) __UpperCAmelCase : Any = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
111
0
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __magic_name__ ( unittest.TestCase ): def __init__( self : str , snake_case__ : List[Any] , snake_case__ : Optional[Any]=1_3 , snake_case__ : List[Any]=7 , snake_case__ : Optional[Any]=True , snake_case__ : Optional[int]=True , snake_case__ : Dict=True , snake_case__ : Optional[Any]=True , snake_case__ : Optional[Any]=9_9 , snake_case__ : Union[str, Any]=3_2 , snake_case__ : Optional[int]=5 , snake_case__ : Optional[int]=4 , snake_case__ : Optional[int]=3_7 , snake_case__ : Optional[Any]="gelu" , snake_case__ : Any=0.1 , snake_case__ : int=0.1 , snake_case__ : int=5_1_2 , snake_case__ : Tuple=1_6 , snake_case__ : Optional[int]=2 , snake_case__ : Any=0.02 , snake_case__ : Dict=4 , ): '''simple docstring''' lowercase :Union[str, Any] = parent lowercase :Tuple = batch_size lowercase :Dict = seq_length lowercase :List[str] = is_training lowercase :int = use_attention_mask lowercase :str = use_token_type_ids lowercase :List[Any] = use_labels lowercase :Tuple = vocab_size lowercase :List[str] = hidden_size lowercase :str = num_hidden_layers lowercase :Optional[Any] = num_attention_heads lowercase :List[str] = intermediate_size lowercase :str = hidden_act lowercase :Optional[Any] = hidden_dropout_prob lowercase :Optional[Any] = attention_probs_dropout_prob lowercase :List[str] = max_position_embeddings lowercase :str = type_vocab_size lowercase :Union[str, Any] = type_sequence_label_size lowercase :Optional[int] = initializer_range lowercase :Tuple = num_choices def __snake_case ( self : Union[str, Any] ): '''simple docstring''' lowercase :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase :Optional[Any] = None if self.use_attention_mask: lowercase :Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowercase :List[Any] = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=snake_case__ , ) return config, input_ids, attention_mask def __snake_case ( self : str ): '''simple docstring''' lowercase :int = self.prepare_config_and_inputs() lowercase , lowercase , lowercase :Union[str, Any] = config_and_inputs lowercase :Any = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class __magic_name__ ( __UpperCAmelCase , unittest.TestCase ): __A : Tuple = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __snake_case ( self : Optional[Any] ): '''simple docstring''' lowercase :int = FlaxDistilBertModelTester(self ) @slow def __snake_case ( self : Dict ): '''simple docstring''' for model_class_name in self.all_model_classes: lowercase :List[Any] = model_class_name.from_pretrained('''distilbert-base-uncased''' ) lowercase :Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case__ ) @require_flax class __magic_name__ ( unittest.TestCase ): @slow def __snake_case ( self : Dict ): '''simple docstring''' lowercase :Dict = FlaxDistilBertModel.from_pretrained('''distilbert-base-uncased''' ) lowercase :int = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) lowercase :List[str] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowercase :str = model(snake_case__ , attention_mask=snake_case__ )[0] lowercase :int = (1, 1_1, 7_6_8) self.assertEqual(output.shape , snake_case__ ) lowercase :Tuple = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , snake_case__ , atol=1e-4 ) )
172
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def lowerCamelCase (a_ :List[Any]) -> Any: lowercase :Tuple = 384 lowercase :int = 7 if "tiny" in model_name: lowercase :Union[str, Any] = 96 lowercase :Any = (2, 2, 6, 2) lowercase :Any = (3, 6, 12, 24) elif "small" in model_name: lowercase :int = 96 lowercase :Any = (2, 2, 18, 2) lowercase :Optional[int] = (3, 6, 12, 24) elif "base" in model_name: lowercase :Optional[int] = 128 lowercase :Any = (2, 2, 18, 2) lowercase :Dict = (4, 8, 16, 32) lowercase :Tuple = 12 lowercase :List[Any] = 512 elif "large" in model_name: lowercase :List[str] = 192 lowercase :Optional[int] = (2, 2, 18, 2) lowercase :Any = (6, 12, 24, 48) lowercase :Any = 12 lowercase :str = 768 # set label information lowercase :Optional[int] = 150 lowercase :str = '''huggingface/label-files''' lowercase :List[str] = '''ade20k-id2label.json''' lowercase :List[str] = json.load(open(hf_hub_download(a_ , a_ , repo_type='''dataset''') , '''r''')) lowercase :Union[str, Any] = {int(a_): v for k, v in idalabel.items()} lowercase :Dict = {v: k for k, v in idalabel.items()} lowercase :List[str] = SwinConfig( embed_dim=a_ , depths=a_ , num_heads=a_ , window_size=a_ , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] , ) lowercase :List[str] = UperNetConfig( backbone_config=a_ , auxiliary_in_channels=a_ , num_labels=a_ , idalabel=a_ , labelaid=a_ , ) return config def lowerCamelCase (a_ :Any) -> Optional[Any]: lowercase :Any = [] # fmt: off # stem rename_keys.append(('''backbone.patch_embed.projection.weight''', '''backbone.embeddings.patch_embeddings.projection.weight''')) rename_keys.append(('''backbone.patch_embed.projection.bias''', '''backbone.embeddings.patch_embeddings.projection.bias''')) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''backbone.embeddings.norm.weight''')) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''backbone.embeddings.norm.bias''')) # stages for i in range(len(config.backbone_config.depths)): for j in range(config.backbone_config.depths[i]): rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.norm1.weight""", F"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.norm1.bias""", F"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table""", F"""backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index""", F"""backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight""", F"""backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias""", F"""backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.norm2.weight""", F"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.norm2.bias""", F"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight""", F"""backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias""", F"""backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight""", F"""backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight""")) rename_keys.append((F"""backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias""", F"""backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias""")) if i < 3: rename_keys.append((F"""backbone.stages.{i}.downsample.reduction.weight""", F"""backbone.encoder.layers.{i}.downsample.reduction.weight""")) rename_keys.append((F"""backbone.stages.{i}.downsample.norm.weight""", F"""backbone.encoder.layers.{i}.downsample.norm.weight""")) rename_keys.append((F"""backbone.stages.{i}.downsample.norm.bias""", F"""backbone.encoder.layers.{i}.downsample.norm.bias""")) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""")) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""")) # decode head rename_keys.extend( [ ('''decode_head.conv_seg.weight''', '''decode_head.classifier.weight'''), ('''decode_head.conv_seg.bias''', '''decode_head.classifier.bias'''), ('''auxiliary_head.conv_seg.weight''', '''auxiliary_head.classifier.weight'''), ('''auxiliary_head.conv_seg.bias''', '''auxiliary_head.classifier.bias'''), ]) # fmt: on return rename_keys def lowerCamelCase (a_ :int , a_ :str , a_ :Union[str, Any]) -> int: lowercase :Any = dct.pop(a_) lowercase :Tuple = val def lowerCamelCase (a_ :Optional[int] , a_ :int) -> Optional[int]: lowercase :int = [int(backbone_config.embed_dim * 2**i) for i in range(len(backbone_config.depths))] for i in range(len(backbone_config.depths)): lowercase :List[str] = num_features[i] for j in range(backbone_config.depths[i]): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) lowercase :Optional[Any] = state_dict.pop(F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight""") lowercase :int = state_dict.pop(F"""backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias""") # next, add query, keys and values (in that order) to the state dict lowercase :Optional[Any] = in_proj_weight[:dim, :] lowercase :str = in_proj_bias[: dim] lowercase :Optional[Any] = in_proj_weight[ dim : dim * 2, : ] lowercase :Optional[int] = in_proj_bias[ dim : dim * 2 ] lowercase :List[str] = in_proj_weight[ -dim :, : ] lowercase :Any = in_proj_bias[-dim :] # fmt: on def lowerCamelCase (a_ :Tuple) -> Any: lowercase , lowercase :Tuple = x.shape lowercase :str = x.reshape(a_ , 4 , in_channel // 4) lowercase :str = x[:, [0, 2, 1, 3], :].transpose(1 , 2).reshape(a_ , a_) return x def lowerCamelCase (a_ :Optional[int]) -> Tuple: lowercase , lowercase :Any = x.shape lowercase :Union[str, Any] = x.reshape(a_ , in_channel // 4 , 4) lowercase :int = x[:, :, [0, 2, 1, 3]].transpose(1 , 2).reshape(a_ , a_) return x def lowerCamelCase (a_ :Any) -> Union[str, Any]: lowercase :int = x.shape[0] lowercase :Any = x.reshape(4 , in_channel // 4) lowercase :Any = x[[0, 2, 1, 3], :].transpose(0 , 1).reshape(a_) return x def lowerCamelCase (a_ :str) -> Union[str, Any]: lowercase :List[str] = x.shape[0] lowercase :Optional[Any] = x.reshape(in_channel // 4 , 4) lowercase :List[Any] = x[:, [0, 2, 1, 3]].transpose(0 , 1).reshape(a_) return x def lowerCamelCase (a_ :int , a_ :Optional[int] , a_ :List[Any]) -> Dict: lowercase :Dict = { '''upernet-swin-tiny''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth''', '''upernet-swin-small''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth''', '''upernet-swin-base''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth''', '''upernet-swin-large''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth''', } lowercase :List[Any] = model_name_to_url[model_name] lowercase :Tuple = torch.hub.load_state_dict_from_url(a_ , map_location='''cpu''' , file_name=a_)[ '''state_dict''' ] for name, param in state_dict.items(): print(a_ , param.shape) lowercase :Union[str, Any] = get_upernet_config(a_) lowercase :int = UperNetForSemanticSegmentation(a_) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): lowercase :Union[str, Any] = state_dict.pop(a_) if "bn" in key: lowercase :int = key.replace('''bn''' , '''batch_norm''') lowercase :str = val # rename keys lowercase :List[str] = create_rename_keys(a_) for src, dest in rename_keys: rename_key(a_ , a_ , a_) read_in_q_k_v(a_ , config.backbone_config) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: lowercase :str = reverse_correct_unfold_reduction_order(a_) if "norm" in key: lowercase :int = reverse_correct_unfold_norm_order(a_) model.load_state_dict(a_) # verify on image lowercase :Dict = '''https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg''' lowercase :str = Image.open(requests.get(a_ , stream=a_).raw).convert('''RGB''') lowercase :Union[str, Any] = SegformerImageProcessor() lowercase :int = processor(a_ , return_tensors='''pt''').pixel_values with torch.no_grad(): lowercase :Optional[int] = model(a_) lowercase :int = outputs.logits print(logits.shape) print('''First values of logits:''' , logits[0, 0, :3, :3]) # assert values if model_name == "upernet-swin-tiny": lowercase :Union[str, Any] = torch.tensor( [[-7.59_58, -7.59_58, -7.43_02], [-7.59_58, -7.59_58, -7.43_02], [-7.47_97, -7.47_97, -7.30_68]]) elif model_name == "upernet-swin-small": lowercase :List[str] = torch.tensor( [[-7.19_21, -7.19_21, -6.95_32], [-7.19_21, -7.19_21, -6.95_32], [-7.09_08, -7.09_08, -6.85_34]]) elif model_name == "upernet-swin-base": lowercase :Tuple = torch.tensor( [[-6.58_51, -6.58_51, -6.43_30], [-6.58_51, -6.58_51, -6.43_30], [-6.47_63, -6.47_63, -6.32_54]]) elif model_name == "upernet-swin-large": lowercase :str = torch.tensor( [[-7.52_97, -7.52_97, -7.38_02], [-7.52_97, -7.52_97, -7.38_02], [-7.40_44, -7.40_44, -7.25_86]]) print('''Logits:''' , outputs.logits[0, 0, :3, :3]) assert torch.allclose(outputs.logits[0, 0, :3, :3] , a_ , atol=1E-4) print('''Looks ok!''') if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""") model.save_pretrained(a_) print(F"""Saving processor to {pytorch_dump_folder_path}""") processor.save_pretrained(a_) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""") model.push_to_hub(F"""openmmlab/{model_name}""") processor.push_to_hub(F"""openmmlab/{model_name}""") if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-swin-tiny''', type=str, choices=[F"""upernet-swin-{size}""" for size in ['''tiny''', '''small''', '''base''', '''large''']], help='''Name of the Swin + UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) UpperCAmelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
172
1
def UpperCAmelCase ( a_ = 1_0_0_0_0_0_0 ) -> int: """simple docstring""" __A = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , a_ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
15
def UpperCAmelCase ( a_ ) -> list: """simple docstring""" if len(a_ ) <= 1: return [tuple(a_ )] __A = [] def generate(a_ , a_ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , a_ ) for i in range(k - 1 ): if k % 2 == 0: # k is even __A , __A = arr[k - 1], arr[i] else: # k is odd __A , __A = arr[k - 1], arr[0] generate(k - 1 , a_ ) generate(len(a_ ) , a_ ) return res if __name__ == "__main__": SCREAMING_SNAKE_CASE :int = input('Enter numbers separated by a comma:\n').strip() SCREAMING_SNAKE_CASE :Dict = [int(item) for item in user_input.split(',')] print(heaps(arr))
15
1
'''simple docstring''' import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings __lowerCAmelCase = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class __magic_name__ ( _UpperCamelCase ): lowerCAmelCase : bool = field(default=_UpperCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowerCAmelCase : bool = field( default=_UpperCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowerCAmelCase : Optional[int] = field( default=_UpperCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowerCAmelCase : Optional[int] = field( default=_UpperCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowerCAmelCase : Optional[Union[str, Path, GenerationConfig]] = field( default=_UpperCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def __lowercase ( self : List[Any] ): _a : Tuple = super().to_dict() for k, v in d.items(): if isinstance(_UpperCAmelCase ,_UpperCAmelCase ): _a : Tuple = v.to_dict() return d
107
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class __magic_name__ : lowerCAmelCase : str = field( metadata={'help': 'The output directory where the model will be written.'} , ) lowerCAmelCase : str = field( metadata={ 'help': ( 'The encoder model checkpoint for weights initialization.' 'Don\'t set if you want to train an encoder model from scratch.' ) } , ) lowerCAmelCase : str = field( metadata={ 'help': ( 'The decoder model checkpoint for weights initialization.' 'Don\'t set if you want to train a decoder model from scratch.' ) } , ) lowerCAmelCase : Optional[str] = field( default=_UpperCamelCase , metadata={'help': 'Pretrained encoder config name or path if not the same as encoder_model_name'} ) lowerCAmelCase : Optional[str] = field( default=_UpperCamelCase , metadata={'help': 'Pretrained decoder config name or path if not the same as decoder_model_name'} ) def __lowerCamelCase ( ) -> Union[str, Any]: _a : Any = HfArgumentParser((ModelArguments,) ) ((_a) , ) : Dict = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _a : Optional[Any] = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _a : Optional[Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _a : List[str] = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _a : Optional[int] = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _a : List[Any] = True _a : int = True _a : Any = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=lowerCAmelCase_ , decoder_config=lowerCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _a : List[str] = decoder_config.decoder_start_token_id _a : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _a : Tuple = decoder_config.bos_token_id if pad_token_id is None: _a : List[Any] = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _a : Any = decoder_config.eos_token_id _a : Tuple = decoder_start_token_id _a : Any = pad_token_id _a : Dict = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _a : Dict = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _a : int = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
107
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Union[str, Any] = logging.get_logger(__name__) __A : List[Any] = { '''weiweishi/roc-bert-base-zh''': '''https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json''', } class A_ (snake_case_ ): UpperCAmelCase__ = '''roc_bert''' def __init__( self , _A=3_0_5_2_2 , _A=7_6_8 , _A=1_2 , _A=1_2 , _A=3_0_7_2 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_2 , _A=2 , _A=0.02 , _A=1E-12 , _A=True , _A=0 , _A="absolute" , _A=None , _A=True , _A=True , _A=7_6_8 , _A=9_1_0 , _A=5_1_2 , _A=2_4_8_5_8 , _A=True , **_A , ): '''simple docstring''' UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = type_vocab_size UpperCAmelCase = layer_norm_eps UpperCAmelCase = use_cache UpperCAmelCase = enable_pronunciation UpperCAmelCase = enable_shape UpperCAmelCase = pronunciation_embed_dim UpperCAmelCase = pronunciation_vocab_size UpperCAmelCase = shape_embed_dim UpperCAmelCase = shape_vocab_size UpperCAmelCase = concat_input UpperCAmelCase = position_embedding_type UpperCAmelCase = classifier_dropout super().__init__(pad_token_id=__snake_case , **__snake_case )
273
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class UpperCAmelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = StableDiffusionControlNetImgaImgPipeline _lowercase: Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) _lowercase: Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self : List[str] ) -> List[str]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Any , __snake_case : str , __snake_case : Any=0 ) -> str: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ) _lowerCAmelCase = floats_tensor(control_image.shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : Optional[int] ) -> List[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = StableDiffusionControlNetImgaImgPipeline _lowercase: Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Any = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__snake_case : Optional[Any] ): if isinstance(__snake_case , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = MultiControlNetModel([controlneta, controlneta] ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Tuple , __snake_case : int , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), ] _lowerCAmelCase = floats_tensor(control_image[0].shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : List[str] ) -> Dict: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) _lowerCAmelCase = 10.0 _lowerCAmelCase = 4 _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def lowercase__ ( self : int ) -> str: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Optional[Any] ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : int ) -> str: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__snake_case ) except NotImplementedError: pass @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Union[str, Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) _lowerCAmelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__snake_case , controlnet=__snake_case ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase = """evil space-punk bird""" _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = pipe( __snake_case , __snake_case , control_image=__snake_case , generator=__snake_case , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) _lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
70
0
'''simple docstring''' def __UpperCamelCase ( _UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = 0 # if input_string is "aba" than new_input_string become "a|b|a" __UpperCAmelCase : Any = "" __UpperCAmelCase : List[Any] = "" # append each character + "|" in new_string for range(0, length-1) for i in input_string[: len(_UpperCAmelCase ) - 1]: new_input_string += i + "|" # append last character new_input_string += input_string[-1] # we will store the starting and ending of previous furthest ending palindromic # substring __UpperCAmelCase : Union[str, Any] = 0, 0 # length[i] shows the length of palindromic substring with center i __UpperCAmelCase : Dict = [1 for i in range(len(_UpperCAmelCase ) )] # for each character in new_string find corresponding palindromic string __UpperCAmelCase : str = 0 for j in range(len(_UpperCAmelCase ) ): __UpperCAmelCase : str = 1 if j > r else min(length[l + r - j] // 2, r - j + 1 ) while ( j - k >= 0 and j + k < len(_UpperCAmelCase ) and new_input_string[k + j] == new_input_string[j - k] ): k += 1 __UpperCAmelCase : Optional[int] = 2 * k - 1 # does this string is ending after the previously explored end (that is r) ? # if yes the update the new r to the last index of this if j + k - 1 > r: __UpperCAmelCase : Tuple = j - k + 1 # noqa: E741 __UpperCAmelCase : int = j + k - 1 # update max_length and start position if max_length < length[j]: __UpperCAmelCase : List[str] = length[j] __UpperCAmelCase : Dict = j # create that string __UpperCAmelCase : List[str] = new_input_string[start - max_length // 2 : start + max_length // 2 + 1] for i in s: if i != "|": output_string += i return output_string if __name__ == "__main__": import doctest doctest.testmod()
360
'''simple docstring''' from datetime import datetime as dt import os from github import Github lowerCAmelCase__ : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def __UpperCamelCase ( ): __UpperCAmelCase : Optional[int] = Github(os.environ["GITHUB_TOKEN"] ) __UpperCAmelCase : Union[str, Any] = g.get_repo("huggingface/transformers" ) __UpperCAmelCase : Union[str, Any] = repo.get_issues(state="open" ) for issue in open_issues: __UpperCAmelCase : int = sorted([comment for comment in issue.get_comments()], key=lambda _UpperCAmelCase : i.created_at, reverse=_UpperCAmelCase ) __UpperCAmelCase : Any = comments[0] if len(_UpperCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="closed" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
37
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() snake_case_ = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) snake_case_ = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def lowerCamelCase__ ( snake_case_ : List[str] , snake_case_ : Dict , snake_case_ : Optional[Any] ) -> Optional[Any]: __snake_case = state_dict.pop(__lowerCamelCase ) __snake_case = val def lowerCamelCase__ ( snake_case_ : Tuple ) -> int: __snake_case = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: __snake_case = key.replace('''backbone.0.body''' , '''backbone.conv_encoder.model''' ) __snake_case = value else: __snake_case = value return new_state_dict def lowerCamelCase__ ( snake_case_ : Dict , snake_case_ : Optional[int]=False ) -> Tuple: __snake_case = "" if is_panoptic: __snake_case = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) __snake_case = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) __snake_case = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __snake_case = in_proj_weight[:256, :] __snake_case = in_proj_bias[:256] __snake_case = in_proj_weight[256:512, :] __snake_case = in_proj_bias[256:512] __snake_case = in_proj_weight[-256:, :] __snake_case = in_proj_bias[-256:] def lowerCamelCase__ ( ) -> Optional[Any]: __snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int ) -> List[Any]: __snake_case = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: __snake_case = "resnet101" if "dc5" in model_name: __snake_case = True __snake_case = "panoptic" in model_name if is_panoptic: __snake_case = 250 else: __snake_case = 91 __snake_case = "huggingface/label-files" __snake_case = "coco-detection-id2label.json" __snake_case = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type='''dataset''' ) , '''r''' ) ) __snake_case = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case = idalabel __snake_case = {v: k for k, v in idalabel.items()} # load image processor __snake_case = "coco_panoptic" if is_panoptic else "coco_detection" __snake_case = ConditionalDetrImageProcessor(format=__lowerCamelCase ) # prepare image __snake_case = prepare_img() __snake_case = image_processor(images=__lowerCamelCase , return_tensors='''pt''' ) __snake_case = encoding["pixel_values"] logger.info(f"""Converting model {model_name}...""" ) # load original model from torch hub __snake_case = torch.hub.load('''DeppMeng/ConditionalDETR''' , __lowerCamelCase , pretrained=__lowerCamelCase ).eval() __snake_case = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: __snake_case = "conditional_detr." + src rename_key(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) __snake_case = rename_backbone_keys(__lowerCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(__lowerCamelCase , is_panoptic=__lowerCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them __snake_case = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('''conditional_detr''' ) and not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ) ): __snake_case = state_dict.pop(__lowerCamelCase ) __snake_case = val elif "class_labels_classifier" in key or "bbox_predictor" in key: __snake_case = state_dict.pop(__lowerCamelCase ) __snake_case = val elif key.startswith('''bbox_attention''' ) or key.startswith('''mask_head''' ): continue else: __snake_case = state_dict.pop(__lowerCamelCase ) __snake_case = val else: if not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ): __snake_case = state_dict.pop(__lowerCamelCase ) __snake_case = val # finally, create HuggingFace model and load state dict __snake_case = ConditionalDetrForSegmentation(__lowerCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(__lowerCamelCase ) model.load_state_dict(__lowerCamelCase ) model.eval() model.push_to_hub(repo_id=__lowerCamelCase , organization='''DepuMeng''' , commit_message='''Add model''' ) # verify our conversion __snake_case = conditional_detr(__lowerCamelCase ) __snake_case = model(__lowerCamelCase ) assert torch.allclose(outputs.logits , original_outputs['''pred_logits'''] , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs['''pred_boxes'''] , atol=1e-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['''pred_masks'''] , atol=1e-4 ) # Save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) model.save_pretrained(__lowerCamelCase ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) snake_case_ = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
24
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : str = checkpoint UpperCAmelCase_ : int = {} UpperCAmelCase_ : Optional[int] = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ : List[str] = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ : Optional[int] = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ : List[Any] = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ : Union[str, Any] = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ : Any = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ : int = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ : Any = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ : List[Any] = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ : str = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ : List[str] = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ : List[Any] = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ : Optional[Any] = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ : Dict = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ : Optional[int] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } for i in range(__lowerCamelCase ): UpperCAmelCase_ : Any = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : Dict = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ : Dict = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ : List[str] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : List[Any] = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ : Tuple = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : str = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ : List[Any] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ : Union[str, Any] = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : int = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) for i in range(__lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = num_up_blocks - 1 - i UpperCAmelCase_ : Any = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : str = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ : Optional[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ : Dict = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : List[str] = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ : List[Any] = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : str = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ : Tuple = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Tuple = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ : Any = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) return new_checkpoint def __a ( __lowerCamelCase, __lowerCamelCase, ): # Only support V1 UpperCAmelCase_ : List[str] = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ : List[Any] = io.BytesIO(r.content ) UpperCAmelCase_ : Any = OmegaConf.load(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = 512 UpperCAmelCase_ : Optional[Any] = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ : int = {} with safe_open(__lowerCamelCase, framework="pt", device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ : Tuple = f.get_tensor(__lowerCamelCase ) else: UpperCAmelCase_ : Any = torch.load(__lowerCamelCase, map_location=__lowerCamelCase )["state_dict"] # Convert the VAE model. UpperCAmelCase_ : Dict = create_vae_diffusers_config(__lowerCamelCase, image_size=__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = custom_convert_ldm_vae_checkpoint(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : int = AutoencoderKL(**__lowerCamelCase ) vae.load_state_dict(__lowerCamelCase ) vae.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument('--vae_pt_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') _a = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
61
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a__ : Union[str, Any] = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Dict = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[str] = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Dict = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys a__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
352
'''simple docstring''' # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar a__ : int = TypeVar('T') class UpperCAmelCase__ ( Generic[T]): def __init__( self , lowercase = True ) -> None: __UpperCamelCase = {} # dictionary of lists __UpperCamelCase = directed def __lowerCamelCase ( self , lowercase , lowercase ) -> GraphAdjacencyList[T]: if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) self.adj_list[destination_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) __UpperCamelCase = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowercase ) __UpperCamelCase = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: __UpperCamelCase = [destination_vertex] __UpperCamelCase = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) __UpperCamelCase = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: __UpperCamelCase = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: __UpperCamelCase = [destination_vertex] __UpperCamelCase = [] return self def __repr__( self ) -> str: return pformat(self.adj_list )
243
0
"""simple docstring""" from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __A = logging.get_logger(__name__) class lowerCamelCase__ ( lowerCamelCase_ ): a__ : Dict = ["""audio_values""", """audio_mask"""] def __init__( self , SCREAMING_SNAKE_CASE=2_048 , SCREAMING_SNAKE_CASE=1 , SCREAMING_SNAKE_CASE=[16, 16] , SCREAMING_SNAKE_CASE=128 , SCREAMING_SNAKE_CASE=44_100 , SCREAMING_SNAKE_CASE=86 , SCREAMING_SNAKE_CASE=2_048 , SCREAMING_SNAKE_CASE=0.0 , **SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__( feature_size=SCREAMING_SNAKE_CASE , sampling_rate=SCREAMING_SNAKE_CASE , padding_value=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) snake_case : Any = spectrogram_length snake_case : Dict = num_channels snake_case : Union[str, Any] = patch_size snake_case : str = feature_size // self.patch_size[1] snake_case : List[Any] = n_fft snake_case : Dict = sampling_rate // hop_length_to_sampling_rate snake_case : Optional[int] = sampling_rate snake_case : List[str] = padding_value snake_case : List[str] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=SCREAMING_SNAKE_CASE , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=SCREAMING_SNAKE_CASE , norm="slaney" , mel_scale="slaney" , ).T def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case : List[Any] = spectrogram( SCREAMING_SNAKE_CASE , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) snake_case : Optional[Any] = log_spec[:, :-1] snake_case : str = log_spec - 20.0 snake_case : int = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , **SCREAMING_SNAKE_CASE , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' F''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) snake_case : Optional[int] = isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) snake_case : List[str] = is_batched_numpy or ( isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: snake_case : str = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ): snake_case : List[str] = np.asarray(SCREAMING_SNAKE_CASE , dtype=np.floataa ) elif isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): snake_case : Dict = raw_speech.astype(np.floataa ) # always return batch if not is_batched: snake_case : Union[str, Any] = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis snake_case : Optional[Any] = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , SCREAMING_SNAKE_CASE ): snake_case : str = [np.asarray(SCREAMING_SNAKE_CASE , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask snake_case : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: snake_case : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] snake_case : Optional[Any] = np.array(SCREAMING_SNAKE_CASE ).astype(np.floataa ) # convert into correct format for padding snake_case : List[Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch snake_case : Optional[Any] = np.ones([len(SCREAMING_SNAKE_CASE ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) snake_case : Optional[int] = padded_audio_features * self.padding_value for i in range(len(SCREAMING_SNAKE_CASE ) ): snake_case : Union[str, Any] = audio_features[i] snake_case : str = feature # return as BatchFeature if return_attention_mask: snake_case : Tuple = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: snake_case : int = {"audio_values": padded_audio_features} snake_case : int = BatchFeature(data=SCREAMING_SNAKE_CASE , tensor_type=SCREAMING_SNAKE_CASE ) return encoded_inputs
148
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __A = logging.get_logger(__name__) if is_vision_available(): import PIL class lowerCamelCase__ ( lowerCamelCase_ ): a__ : Union[str, Any] = ["""pixel_values"""] def __init__( self , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = PILImageResampling.BICUBIC , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = 1 / 255 , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , **SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE ) snake_case : int = size if size is not None else {"shortest_edge": 224} snake_case : int = get_size_dict(SCREAMING_SNAKE_CASE , default_to_square=SCREAMING_SNAKE_CASE ) snake_case : List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} snake_case : Tuple = get_size_dict(SCREAMING_SNAKE_CASE , default_to_square=SCREAMING_SNAKE_CASE , param_name="crop_size" ) snake_case : Dict = do_resize snake_case : Optional[int] = size snake_case : int = resample snake_case : Union[str, Any] = do_center_crop snake_case : Dict = crop_size snake_case : Dict = do_rescale snake_case : Any = rescale_factor snake_case : Tuple = do_normalize snake_case : int = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case : Tuple = image_std if image_std is not None else OPENAI_CLIP_STD snake_case : Tuple = do_convert_rgb def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = PILImageResampling.BICUBIC , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ): """simple docstring""" snake_case : List[Any] = get_size_dict(SCREAMING_SNAKE_CASE , default_to_square=SCREAMING_SNAKE_CASE ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) snake_case : Dict = get_resize_output_image_size(SCREAMING_SNAKE_CASE , size=size["shortest_edge"] , default_to_square=SCREAMING_SNAKE_CASE ) return resize(SCREAMING_SNAKE_CASE , size=SCREAMING_SNAKE_CASE , resample=SCREAMING_SNAKE_CASE , data_format=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ): """simple docstring""" snake_case : Tuple = get_size_dict(SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(SCREAMING_SNAKE_CASE , size=(size["height"], size["width"]) , data_format=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ): """simple docstring""" return rescale(SCREAMING_SNAKE_CASE , scale=SCREAMING_SNAKE_CASE , data_format=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ): """simple docstring""" return normalize(SCREAMING_SNAKE_CASE , mean=SCREAMING_SNAKE_CASE , std=SCREAMING_SNAKE_CASE , data_format=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE , ): """simple docstring""" snake_case : int = do_resize if do_resize is not None else self.do_resize snake_case : List[str] = size if size is not None else self.size snake_case : Dict = get_size_dict(SCREAMING_SNAKE_CASE , param_name="size" , default_to_square=SCREAMING_SNAKE_CASE ) snake_case : Optional[Any] = resample if resample is not None else self.resample snake_case : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case : Optional[int] = crop_size if crop_size is not None else self.crop_size snake_case : Union[str, Any] = get_size_dict(SCREAMING_SNAKE_CASE , param_name="crop_size" , default_to_square=SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale snake_case : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case : Union[str, Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case : List[str] = image_mean if image_mean is not None else self.image_mean snake_case : Optional[int] = image_std if image_std is not None else self.image_std snake_case : Optional[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case : List[Any] = make_list_of_images(SCREAMING_SNAKE_CASE ) if not valid_images(SCREAMING_SNAKE_CASE ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case : Optional[int] = [convert_to_rgb(SCREAMING_SNAKE_CASE ) for image in images] # All transformations expect numpy arrays. snake_case : List[str] = [to_numpy_array(SCREAMING_SNAKE_CASE ) for image in images] if do_resize: snake_case : Optional[Any] = [self.resize(image=SCREAMING_SNAKE_CASE , size=SCREAMING_SNAKE_CASE , resample=SCREAMING_SNAKE_CASE ) for image in images] if do_center_crop: snake_case : int = [self.center_crop(image=SCREAMING_SNAKE_CASE , size=SCREAMING_SNAKE_CASE ) for image in images] if do_rescale: snake_case : str = [self.rescale(image=SCREAMING_SNAKE_CASE , scale=SCREAMING_SNAKE_CASE ) for image in images] if do_normalize: snake_case : Optional[int] = [self.normalize(image=SCREAMING_SNAKE_CASE , mean=SCREAMING_SNAKE_CASE , std=SCREAMING_SNAKE_CASE ) for image in images] snake_case : Optional[int] = [to_channel_dimension_format(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for image in images] snake_case : Tuple = {"pixel_values": images} return BatchFeature(data=SCREAMING_SNAKE_CASE , tensor_type=SCREAMING_SNAKE_CASE )
148
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCamelCase__ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case : Union[str, Any] = ["image_processor", "tokenizer"] __snake_case : Dict = "CLIPImageProcessor" __snake_case : List[Any] = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self : Optional[Any] ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : List[Any]=None ,**lowerCamelCase__ : str ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,lowerCamelCase__ ,) SCREAMING_SNAKE_CASE = kwargs.pop("""feature_extractor""" ) SCREAMING_SNAKE_CASE = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(lowerCamelCase__ ,lowerCamelCase__ ) def __call__( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any=None ,lowerCamelCase__ : Dict=None ,**lowerCamelCase__ : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: SCREAMING_SNAKE_CASE = self.tokenizer(lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) if images is not None: SCREAMING_SNAKE_CASE = self.image_processor(lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) if text is not None and images is not None: SCREAMING_SNAKE_CASE = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowerCamelCase__ ) ,tensor_type=lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ,*lowerCamelCase__ : Optional[Any] ,**lowerCamelCase__ : Union[str, Any] ) -> int: '''simple docstring''' return self.tokenizer.batch_decode(*lowerCamelCase__ ,**lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ,*lowerCamelCase__ : Optional[int] ,**lowerCamelCase__ : Union[str, Any] ) -> List[Any]: '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase__ ,**lowerCamelCase__ ) @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
193
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[int] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any]=3 ,lowerCamelCase__ : List[str]=32 ,lowerCamelCase__ : List[Any]=3 ,lowerCamelCase__ : str=10 ,lowerCamelCase__ : Any=[10, 20, 30, 40] ,lowerCamelCase__ : Optional[Any]=[1, 1, 2, 1] ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Tuple="relu" ,lowerCamelCase__ : Dict=3 ,lowerCamelCase__ : Optional[int]=None ,) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = embeddings_size SCREAMING_SNAKE_CASE = hidden_sizes SCREAMING_SNAKE_CASE = depths SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = num_labels SCREAMING_SNAKE_CASE = scope SCREAMING_SNAKE_CASE = len(lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values def SCREAMING_SNAKE_CASE__ ( self : str ) -> Optional[int]: '''simple docstring''' return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = FlaxRegNetModel(config=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = model(lowerCamelCase__ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[Any] ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = self.num_labels SCREAMING_SNAKE_CASE = FlaxRegNetForImageClassification(config=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"""pixel_values""": pixel_values} return config, inputs_dict @require_flax class UpperCamelCase__ ( lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' __snake_case : Union[str, Any] = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () __snake_case : Tuple = False __snake_case : int = False __snake_case : Tuple = False def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> None: '''simple docstring''' SCREAMING_SNAKE_CASE = FlaxRegNetModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self ,config_class=lowerCamelCase__ ,has_text_modality=lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> Optional[int]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE__ ( self : int ) -> List[str]: '''simple docstring''' return def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> Any: '''simple docstring''' pass @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> Any: '''simple docstring''' pass def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> int: '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ): SCREAMING_SNAKE_CASE = model_class(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) ) SCREAMING_SNAKE_CASE = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states SCREAMING_SNAKE_CASE = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) ,expected_num_stages + 1 ) SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) SCREAMING_SNAKE_CASE = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ : Dict ,**lowerCamelCase__ : Optional[Any] ): return model(pixel_values=lowerCamelCase__ ,**lowerCamelCase__ ) with self.subTest("""JIT Enabled""" ): SCREAMING_SNAKE_CASE = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): SCREAMING_SNAKE_CASE = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) ,len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_flax class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self : int ) -> str: '''simple docstring''' return AutoImageProcessor.from_pretrained("""facebook/regnet-y-040""" ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = FlaxRegNetForImageClassification.from_pretrained("""facebook/regnet-y-040""" ) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=lowerCamelCase__ ,return_tensors="""np""" ) SCREAMING_SNAKE_CASE = model(**lowerCamelCase__ ) # verify the logits SCREAMING_SNAKE_CASE = (1, 1000) self.assertEqual(outputs.logits.shape ,lowerCamelCase__ ) SCREAMING_SNAKE_CASE = jnp.array([-0.4180, -1.5051, -3.4836] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] ,lowerCamelCase__ ,atol=1e-4 ) )
193
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger lowercase__ : Optional[Any] = get_logger(__name__) class lowercase_ ( enum.Enum ): """simple docstring""" UpperCAmelCase_ : Optional[int] = """all_checks""" UpperCAmelCase_ : List[str] = """basic_checks""" UpperCAmelCase_ : Union[str, Any] = """no_checks""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ , snake_case__=None ) -> Union[str, Any]: if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(snake_case__ ) - set(snake_case__ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(snake_case__ ) - set(snake_case__ ) ) ) if len(set(snake_case__ ) - set(snake_case__ ) ) > 0: raise UnexpectedDownloadedFile(str(set(snake_case__ ) - set(snake_case__ ) ) ) lowerCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] lowerCAmelCase = ''' for ''' + verification_name if verification_name is not None else '''''' if len(snake_case__ ) > 0: raise NonMatchingChecksumError( f"Checksums didn't match{for_verification_name}:\n" f"{bad_urls}\n" '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" class lowercase_ ( UpperCamelCase_ ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ ) -> Tuple: if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(snake_case__ ) - set(snake_case__ ) ) > 0: raise ExpectedMoreSplits(str(set(snake_case__ ) - set(snake_case__ ) ) ) if len(set(snake_case__ ) - set(snake_case__ ) ) > 0: raise UnexpectedSplits(str(set(snake_case__ ) - set(snake_case__ ) ) ) lowerCAmelCase = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(snake_case__ ) > 0: raise NonMatchingSplitsSizesError(str(snake_case__ ) ) logger.info('''All the splits matched successfully.''' ) def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ = True ) -> dict: if record_checksum: lowerCAmelCase = shaaaa() with open(snake_case__ , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 2_0 ) , B'''''' ): m.update(snake_case__ ) lowerCAmelCase = m.hexdigest() else: lowerCAmelCase = None return {"num_bytes": os.path.getsize(snake_case__ ), "checksum": checksum} def SCREAMING_SNAKE_CASE_ ( snake_case__ ) -> Dict: if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
338
import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( '''split_dict''' , [ SplitDict(), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_3_3_7 , num_examples=4_2 , dataset_name='''my_dataset''' )} ), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_3_3_7 , num_examples=4_2 )} ), SplitDict({'''train''': SplitInfo()} ), ] , ) def SCREAMING_SNAKE_CASE_ ( snake_case__ ) -> Union[str, Any]: lowerCAmelCase = split_dict._to_yaml_list() assert len(snake_case__ ) == len(snake_case__ ) lowerCAmelCase = SplitDict._from_yaml_list(snake_case__ ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump lowerCAmelCase = None # the split name of split_dict takes over the name of the split info object lowerCAmelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( '''split_info''' , [SplitInfo(), SplitInfo(dataset_name=snake_case__ ), SplitInfo(dataset_name='''my_dataset''' )] ) def SCREAMING_SNAKE_CASE_ ( snake_case__ ) -> Optional[int]: # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files lowerCAmelCase = asdict(SplitDict({'''train''': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
338
1
'''simple docstring''' from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : List[str]=True ): '''simple docstring''' model.train() UpperCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = F.mse_loss(SCREAMING_SNAKE_CASE__ , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[Any]=False ): '''simple docstring''' set_seed(42 ) UpperCAmelCase__ = RegressionModel() UpperCAmelCase__ = deepcopy(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = RegressionDataset(length=80 ) UpperCAmelCase__ = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=16 ) model.to(accelerator.device ) if sched: UpperCAmelCase__ = AdamW(params=model.parameters() , lr=1e-3 ) UpperCAmelCase__ = AdamW(params=ddp_model.parameters() , lr=1e-3 ) UpperCAmelCase__ = LambdaLR(SCREAMING_SNAKE_CASE__ , lr_lambda=lambda SCREAMING_SNAKE_CASE__ : epoch**0.65 ) UpperCAmelCase__ = LambdaLR(SCREAMING_SNAKE_CASE__ , lr_lambda=lambda SCREAMING_SNAKE_CASE__ : epoch**0.65 ) # Make a copy of `model` if sched: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = accelerator.prepare(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: UpperCAmelCase__ , UpperCAmelCase__ = accelerator.prepare(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = get_training_setup(SCREAMING_SNAKE_CASE__ ) # Use a single batch UpperCAmelCase__ , UpperCAmelCase__ = next(iter(SCREAMING_SNAKE_CASE__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCAmelCase__ , UpperCAmelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase__ , UpperCAmelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(SCREAMING_SNAKE_CASE__ ): step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: # Sync grads step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase__ = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE__ ) )] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = get_training_setup(SCREAMING_SNAKE_CASE__ ) # Use a single batch UpperCAmelCase__ , UpperCAmelCase__ = next(iter(SCREAMING_SNAKE_CASE__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCAmelCase__ , UpperCAmelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase__ , UpperCAmelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(SCREAMING_SNAKE_CASE__ ): step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: # Sync grads step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase__ = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE__ ) )] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int=False , SCREAMING_SNAKE_CASE__ : List[str]=False ): '''simple docstring''' UpperCAmelCase__ = Accelerator( split_batches=SCREAMING_SNAKE_CASE__ , dispatch_batches=SCREAMING_SNAKE_CASE__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = get_training_setup(SCREAMING_SNAKE_CASE__ ) for iteration, batch in enumerate(SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ , UpperCAmelCase__ = batch.values() # Gather the distributed inputs and targs for the base model UpperCAmelCase__ , UpperCAmelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase__ , UpperCAmelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Do "gradient accumulation" (noop) with accelerator.accumulate(SCREAMING_SNAKE_CASE__ ): step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(SCREAMING_SNAKE_CASE__ ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase__ = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE__ ) )] GradientState._reset_state() def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : Tuple=False ): '''simple docstring''' UpperCAmelCase__ = Accelerator( split_batches=SCREAMING_SNAKE_CASE__ , dispatch_batches=SCREAMING_SNAKE_CASE__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = get_training_setup(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for iteration, batch in enumerate(SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ , UpperCAmelCase__ = batch.values() # Gather the distributed inputs and targs for the base model UpperCAmelCase__ , UpperCAmelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase__ , UpperCAmelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(SCREAMING_SNAKE_CASE__ )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(SCREAMING_SNAKE_CASE__ ): step_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n''' UpperCAmelCase__ = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(SCREAMING_SNAKE_CASE__ )) if accelerator.num_processes > 1: check_model_parameters(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = Accelerator() UpperCAmelCase__ = RegressionDataset(length=80 ) UpperCAmelCase__ = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=16 ) UpperCAmelCase__ = RegressionDataset(length=96 ) UpperCAmelCase__ = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=16 ) UpperCAmelCase__ , UpperCAmelCase__ = accelerator.prepare(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(SCREAMING_SNAKE_CASE__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(SCREAMING_SNAKE_CASE__ ) if iteration < len(SCREAMING_SNAKE_CASE__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(SCREAMING_SNAKE_CASE__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(SCREAMING_SNAKE_CASE__ ) if batch_num < len(SCREAMING_SNAKE_CASE__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = Accelerator() UpperCAmelCase__ = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(SCREAMING_SNAKE_CASE__ ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(SCREAMING_SNAKE_CASE__ ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""" , """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , """`split_batches=False`, `dispatch_batches=False`**""" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation_with_opt_and_scheduler(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' main() if __name__ == "__main__": main()
61
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : List[str] = """yolos""" def __init__( self : str , _UpperCAmelCase : int=7_68 , _UpperCAmelCase : Dict=12 , _UpperCAmelCase : Union[str, Any]=12 , _UpperCAmelCase : Tuple=30_72 , _UpperCAmelCase : int="gelu" , _UpperCAmelCase : Optional[Any]=0.0 , _UpperCAmelCase : List[str]=0.0 , _UpperCAmelCase : Dict=0.02 , _UpperCAmelCase : Optional[Any]=1E-12 , _UpperCAmelCase : Tuple=[5_12, 8_64] , _UpperCAmelCase : str=16 , _UpperCAmelCase : Optional[int]=3 , _UpperCAmelCase : Tuple=True , _UpperCAmelCase : int=1_00 , _UpperCAmelCase : List[str]=True , _UpperCAmelCase : Dict=False , _UpperCAmelCase : Dict=1 , _UpperCAmelCase : Union[str, Any]=5 , _UpperCAmelCase : Dict=2 , _UpperCAmelCase : int=5 , _UpperCAmelCase : str=2 , _UpperCAmelCase : Tuple=0.1 , **_UpperCAmelCase : List[Any] , ): """simple docstring""" super().__init__(**_UpperCAmelCase ) UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = image_size UpperCAmelCase__ = patch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = qkv_bias UpperCAmelCase__ = num_detection_tokens UpperCAmelCase__ = use_mid_position_embeddings UpperCAmelCase__ = auxiliary_loss # Hungarian matcher UpperCAmelCase__ = class_cost UpperCAmelCase__ = bbox_cost UpperCAmelCase__ = giou_cost # Loss coefficients UpperCAmelCase__ = bbox_loss_coefficient UpperCAmelCase__ = giou_loss_coefficient UpperCAmelCase__ = eos_coefficient class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : List[str] = version.parse("""1.11""" ) @property def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return 1E-4 @property def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" return 12
61
1
"""simple docstring""" # limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( '''pipelines_utils''', '''0.22.0''', '''Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.''', standard_warn=False, stacklevel=3, )
160
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A = { '''configuration_git''': ['''GIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GitConfig''', '''GitVisionConfig'''], '''processing_git''': ['''GitProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''GIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GitForCausalLM''', '''GitModel''', '''GitPreTrainedModel''', '''GitVisionModel''', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
160
1
# using dfs for finding eulerian path traversal def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase=None ) -> Any: snake_case : Union[str, Any] = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: snake_case , snake_case : int = True, True snake_case : List[Any] = dfs(lowercase ,lowercase ,lowercase ,lowercase ) return path def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> Tuple: snake_case : Union[str, Any] = 0 snake_case : Union[str, Any] = -1 for i in range(lowercase ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 snake_case : str = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> List[str]: snake_case : Tuple = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] snake_case , snake_case : Any = check_circuit_or_path(lowercase ,lowercase ) if check == 3: print("""graph is not Eulerian""" ) print("""no path""" ) return snake_case : str = 1 if check == 2: snake_case : Optional[int] = odd_node print("""graph has a Euler path""" ) if check == 1: print("""graph has a Euler cycle""" ) snake_case : Dict = dfs(lowercase ,lowercase ,lowercase ) print(lowercase ) def SCREAMING_SNAKE_CASE__ ( ) -> Optional[Any]: snake_case : Union[str, Any] = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} snake_case : Optional[int] = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} snake_case : Optional[int] = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} snake_case : List[Any] = {1: [2, 3], 2: [1, 3], 3: [1, 2]} snake_case : Tuple = { 1: [], 2: [] # all degree is zero } snake_case : Tuple = 10 check_euler(lowercase ,lowercase ) check_euler(lowercase ,lowercase ) check_euler(lowercase ,lowercase ) check_euler(lowercase ,lowercase ) check_euler(lowercase ,lowercase ) if __name__ == "__main__": main()
176
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase=8 ) -> Optional[int]: snake_case : Dict = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : int = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class __lowercase (UpperCamelCase__ ): """simple docstring""" def __init__( self , A , A , A , ) -> int: super().__init__() self.register_modules( unet=A , scheduler=A , movq=A , ) snake_case : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self , A , A , A , A , A , A ) -> str: if latents is None: snake_case : List[str] = randn_tensor(A , generator=A , device=A , dtype=A ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) snake_case : Union[str, Any] = latents.to(A ) snake_case : Optional[Any] = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self , A=0 ) -> List[str]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) snake_case : Dict = torch.device(f"""cuda:{gpu_id}""" ) snake_case : Tuple = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A , A ) def UpperCAmelCase ( self , A=0 ) -> Optional[Any]: if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) snake_case : int = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : Tuple = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Tuple = cpu_offload_with_hook(A , A , prev_module_hook=A ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self ) -> str: if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(A , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A ) def __call__( self , A , A , A , A = 5_1_2 , A = 5_1_2 , A = 1_0_0 , A = 4.0 , A = 1 , A = None , A = None , A = "pil" , A = True , ) -> Any: snake_case : int = self._execution_device snake_case : Tuple = guidance_scale > 1.0 if isinstance(A , A ): snake_case : str = torch.cat(A , dim=0 ) if isinstance(A , A ): snake_case : int = torch.cat(A , dim=0 ) if isinstance(A , A ): snake_case : Tuple = torch.cat(A , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Optional[int] = image_embeds.repeat_interleave(A , dim=0 ) snake_case : Dict = negative_image_embeds.repeat_interleave(A , dim=0 ) snake_case : Optional[Any] = hint.repeat_interleave(A , dim=0 ) snake_case : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A ) snake_case : Optional[Any] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=A ) self.scheduler.set_timesteps(A , device=A ) snake_case : Tuple = self.scheduler.timesteps snake_case : Tuple = self.movq.config.latent_channels snake_case , snake_case : Optional[int] = downscale_height_and_width(A , A , self.movq_scale_factor ) # create initial latent snake_case : str = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A , A , A , self.scheduler , ) for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance snake_case : str = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Any = {"""image_embeds""": image_embeds, """hint""": hint} snake_case : List[Any] = self.unet( sample=A , timestep=A , encoder_hidden_states=A , added_cond_kwargs=A , return_dict=A , )[0] if do_classifier_free_guidance: snake_case , snake_case : Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Tuple = noise_pred.chunk(2 ) snake_case , snake_case : Optional[Any] = variance_pred.chunk(2 ) snake_case : Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : Dict = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : int = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : Dict = self.scheduler.step( A , A , A , generator=A , )[0] # post-processing snake_case : Tuple = self.movq.decode(A , force_not_quantize=A )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: snake_case : Optional[int] = image * 0.5 + 0.5 snake_case : Union[str, Any] = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : Tuple = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
176
1
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger __A = "<<<<<<< This should probably be modified because it mentions: " __A = "=======\n>>>>>>>\n" __A = [ "TextEncoderConfig", "ByteTextEncoder", "SubwordTextEncoder", "encoder_config", "maybe_build_from_corpus", "manual_dir", ] __A = [ # (pattern, replacement) # Order is important here for some replacements (R"tfds\.core", R"datasets"), (R"tf\.io\.gfile\.GFile", R"open"), (R"tf\.([\w\d]+)", R"datasets.Value('\1')"), (R"tfds\.features\.Text\(\)", R"datasets.Value('string')"), (R"tfds\.features\.Text\(", R"datasets.Value('string'),"), (R"features\s*=\s*tfds.features.FeaturesDict\(", R"features=datasets.Features("), (R"tfds\.features\.FeaturesDict\(", R"dict("), (R"The TensorFlow Datasets Authors", R"The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"), (R"tfds\.", R"datasets."), (R"dl_manager\.manual_dir", R"self.config.data_dir"), (R"self\.builder_config", R"self.config"), ] def lowerCAmelCase_ ( __a ) -> Optional[Any]: """simple docstring""" return ConvertCommand(args.tfds_path , args.datasets_directory ) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @staticmethod def SCREAMING_SNAKE_CASE_ (UpperCAmelCase_ : ArgumentParser) ->Dict: '''simple docstring''' lowerCamelCase__: List[str] =parser.add_parser( "convert" , help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset." , ) train_parser.add_argument( "--tfds_path" , type=UpperCAmelCase_ , required=UpperCAmelCase_ , help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert." , ) train_parser.add_argument( "--datasets_directory" , type=UpperCAmelCase_ , required=UpperCAmelCase_ , help="Path to the HuggingFace Datasets folder.") train_parser.set_defaults(func=UpperCAmelCase_) def __init__(self : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : str , *UpperCAmelCase_ : List[str]) ->int: '''simple docstring''' lowerCamelCase__: str =get_logger("datasets-cli/converting") lowerCamelCase__: Tuple =tfds_path lowerCamelCase__: Union[str, Any] =datasets_directory def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Union[str, Any]: '''simple docstring''' if os.path.isdir(self._tfds_path): lowerCamelCase__: Dict =os.path.abspath(self._tfds_path) elif os.path.isfile(self._tfds_path): lowerCamelCase__: Any =os.path.dirname(self._tfds_path) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path.") lowerCamelCase__: Any =os.path.abspath(self._datasets_directory) self._logger.info(F"""Converting datasets from {abs_tfds_path} to {abs_datasets_path}""") lowerCamelCase__: List[str] =[] lowerCamelCase__: Optional[int] =[] lowerCamelCase__: int ={} if os.path.isdir(self._tfds_path): lowerCamelCase__: Tuple =os.listdir(UpperCAmelCase_) else: lowerCamelCase__: int =[os.path.basename(self._tfds_path)] for f_name in file_names: self._logger.info(F"""Looking at file {f_name}""") lowerCamelCase__: Tuple =os.path.join(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: List[str] =os.path.join(UpperCAmelCase_ , UpperCAmelCase_) if not os.path.isfile(UpperCAmelCase_) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file") continue with open(UpperCAmelCase_ , encoding="utf-8") as f: lowerCamelCase__: Union[str, Any] =f.readlines() lowerCamelCase__: int =[] lowerCamelCase__: Any =False lowerCamelCase__: int =False lowerCamelCase__: int =[] for line in lines: lowerCamelCase__: List[Any] =line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowerCamelCase__: List[Any] ="import datasets\n" elif "import tensorflow" in out_line: # order is important here lowerCamelCase__: Union[str, Any] ="" continue elif "from absl import logging" in out_line: lowerCamelCase__: Tuple ="from datasets import logging\n" elif "getLogger" in out_line: lowerCamelCase__: List[str] =out_line.replace("getLogger" , "get_logger") elif any(expression in out_line for expression in TO_HIGHLIGHT): lowerCamelCase__: str =True lowerCamelCase__: List[Any] =list(filter(lambda UpperCAmelCase_: e in out_line , UpperCAmelCase_)) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(UpperCAmelCase_) + "\n") out_lines.append(UpperCAmelCase_) out_lines.append(UpperCAmelCase_) continue else: for pattern, replacement in TO_CONVERT: lowerCamelCase__: Dict =re.sub(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowerCamelCase__: Any =re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)" , UpperCAmelCase_) tfds_imports.extend(imp.strip() for imp in match.group(1).split(",")) lowerCamelCase__: Any ="from . import " + match.group(1) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F"""Error converting {out_line.strip()}""") if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowerCamelCase__: Optional[int] =True out_lines.append(UpperCAmelCase_) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowerCamelCase__: Tuple =f_name.replace(".py" , "") lowerCamelCase__: Optional[int] =os.path.join(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: Optional[Any] =os.path.join(UpperCAmelCase_ , UpperCAmelCase_) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_) self._logger.info(F"""Adding directory {output_dir}""") imports_to_builder_map.update({imp: output_dir for imp in tfds_imports}) else: # Utilities will be moved at the end utils_files.append(UpperCAmelCase_) if needs_manual_update: with_manual_update.append(UpperCAmelCase_) with open(UpperCAmelCase_ , "w" , encoding="utf-8") as f: f.writelines(UpperCAmelCase_) self._logger.info(F"""Converted in {output_file}""") for utils_file in utils_files: try: lowerCamelCase__: Union[str, Any] =os.path.basename(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =imports_to_builder_map[f_name.replace(".py" , "")] self._logger.info(F"""Moving {dest_folder} to {utils_file}""") shutil.copy(UpperCAmelCase_ , UpperCAmelCase_) except KeyError: self._logger.error(F"""Cannot find destination folder for {utils_file}. Please copy manually.""") if with_manual_update: for file_path in with_manual_update: self._logger.warning( F"""You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'.""")
10
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" if not numbers: return 0 if not isinstance(__snake_case, (list, tuple) ) or not all( isinstance(__snake_case, __snake_case ) for number in numbers ): raise ValueError('''numbers must be an iterable of integers''' ) _UpperCamelCase = _UpperCamelCase = _UpperCamelCase = numbers[0] for i in range(1, len(__snake_case ) ): # update the maximum and minimum subarray products _UpperCamelCase = numbers[i] if number < 0: _UpperCamelCase , _UpperCamelCase = min_till_now, max_till_now _UpperCamelCase = max(__snake_case, max_till_now * number ) _UpperCamelCase = min(__snake_case, min_till_now * number ) # update the maximum product found till now _UpperCamelCase = max(__snake_case, __snake_case ) return max_prod
100
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=False , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Tuple: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> str: '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__a , initializer_range=self.initializer_range , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = BioGptModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a , __a , __a , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = BioGptForCausalLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , *__a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = BioGptModel(config=__a) model.to(__a) model.eval() # create attention mask _UpperCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=__a) _UpperCamelCase = self.seq_length // 2 _UpperCamelCase = 0 # first forward pass _UpperCamelCase , _UpperCamelCase = model(__a , attention_mask=__a).to_tuple() # create hypothetical next token and extent to next_input_ids _UpperCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size) # change a random masked slice from input_ids _UpperCamelCase = ids_tensor((1,) , __a).item() + 1 _UpperCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size).squeeze(-1) _UpperCamelCase = random_other_next_tokens # append to next input_ids and attn_mask _UpperCamelCase = torch.cat([input_ids, next_tokens] , dim=-1) _UpperCamelCase = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=__a)] , dim=1 , ) # get two different outputs _UpperCamelCase = model(__a , attention_mask=__a)['''last_hidden_state'''] _UpperCamelCase = model(__a , past_key_values=__a , attention_mask=__a)['''last_hidden_state'''] # select random slice _UpperCamelCase = ids_tensor((1,) , output_from_past.shape[-1]).item() _UpperCamelCase = output_from_no_past[:, -1, random_slice_idx].detach() _UpperCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__a , __a , atol=1e-3)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , *__a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = BioGptModel(config=__a).to(__a).eval() _UpperCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=__a) # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids _UpperCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size) _UpperCamelCase = ids_tensor((self.batch_size, 3) , 2) # append to next input_ids and _UpperCamelCase = torch.cat([input_ids, next_tokens] , dim=-1) _UpperCamelCase = torch.cat([attention_mask, next_attn_mask] , dim=-1) _UpperCamelCase = model(__a , attention_mask=__a)['''last_hidden_state'''] _UpperCamelCase = model(__a , attention_mask=__a , past_key_values=__a)[ '''last_hidden_state''' ] # select random slice _UpperCamelCase = ids_tensor((1,) , output_from_past.shape[-1]).item() _UpperCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _UpperCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__a , __a , atol=1e-3)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , *__a , __a=False) -> List[Any]: '''simple docstring''' _UpperCamelCase = BioGptForCausalLM(__a) model.to(__a) if gradient_checkpointing: model.gradient_checkpointing_enable() _UpperCamelCase = model(__a , labels=__a) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) result.loss.backward() def UpperCAmelCase ( self , __a , *__a) -> Any: '''simple docstring''' _UpperCamelCase = BioGptModel(__a) _UpperCamelCase = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key]) - model_std) , 0.001) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key]) - 0.0) , 0.01) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , *__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = BioGptForTokenClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowercase__ = (BioGptForCausalLM,) if is_torch_available() else () lowercase__ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = False def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = BioGptModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*__a , gradient_checkpointing=__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*__a) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*__a) @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') model.to(__a) _UpperCamelCase = BioGptTokenizer.from_pretrained('''microsoft/biogpt''') _UpperCamelCase = '''left''' # Define PAD Token = EOS Token = 50256 _UpperCamelCase = tokenizer.eos_token _UpperCamelCase = model.config.eos_token_id # use different length sentences to test batching _UpperCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] _UpperCamelCase = tokenizer(__a , return_tensors='''pt''' , padding=__a) _UpperCamelCase = inputs['''input_ids'''].to(__a) _UpperCamelCase = model.generate( input_ids=__a , attention_mask=inputs['''attention_mask'''].to(__a) , ) _UpperCamelCase = tokenizer(sentences[0] , return_tensors='''pt''').input_ids.to(__a) _UpperCamelCase = model.generate(input_ids=__a) _UpperCamelCase = inputs_non_padded.shape[-1] - inputs['''attention_mask'''][-1].long().sum().cpu().item() _UpperCamelCase = tokenizer(sentences[1] , return_tensors='''pt''').input_ids.to(__a) _UpperCamelCase = model.generate(input_ids=__a , max_length=model.config.max_length - num_paddings) _UpperCamelCase = tokenizer.batch_decode(__a , skip_special_tokens=__a) _UpperCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__a) _UpperCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__a) _UpperCamelCase = [ '''Hello, my dog is a little bit bigger than a little bit.''', '''Today, I have a good idea of how to use the information''', ] self.assertListEqual(__a , __a) self.assertListEqual(__a , [non_padded_sentence, padded_sentence]) @slow def UpperCAmelCase ( self) -> int: '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = BioGptModel.from_pretrained(__a) self.assertIsNotNone(__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = 3 _UpperCamelCase = input_dict['''input_ids'''] _UpperCamelCase = input_ids.ne(1).to(__a) _UpperCamelCase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) _UpperCamelCase = BioGptForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , labels=__a) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = 3 _UpperCamelCase = '''multi_label_classification''' _UpperCamelCase = input_dict['''input_ids'''] _UpperCamelCase = input_ids.ne(1).to(__a) _UpperCamelCase = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) _UpperCamelCase = BioGptForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , labels=__a) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') _UpperCamelCase = torch.tensor([[2, 48_05, 9, 6_56, 21]]) _UpperCamelCase = model(__a)[0] _UpperCamelCase = 4_23_84 _UpperCamelCase = torch.Size((1, 5, vocab_size)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4)) @slow def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = BioGptTokenizer.from_pretrained('''microsoft/biogpt''') _UpperCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') model.to(__a) torch.manual_seed(0) _UpperCamelCase = tokenizer('''COVID-19 is''' , return_tensors='''pt''').to(__a) _UpperCamelCase = model.generate( **__a , min_length=1_00 , max_length=10_24 , num_beams=5 , early_stopping=__a , ) _UpperCamelCase = tokenizer.decode(output_ids[0] , skip_special_tokens=__a) _UpperCamelCase = ( '''COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the''' ''' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and''' ''' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),''' ''' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and''' ''' more than 800,000 deaths.''' ) self.assertEqual(__a , __a)
100
1
"""simple docstring""" import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class a__ ( a_, unittest.TestCase ): __lowerCAmelCase = BlenderbotSmallTokenizer __lowerCAmelCase = False def __magic_name__ ( self ): super().setUp() lowercase : str = ["__start__", "adapt", "act", "ap@@", "te", "__end__", "__unk__"] lowercase : Optional[int] = dict(zip(_a , range(len(_a ) ) ) ) lowercase : int = ["#version: 0.2", "a p", "t e</w>", "ap t</w>", "a d", "ad apt</w>", "a c", "ac t</w>", ""] lowercase : Dict = {"unk_token": "__unk__", "bos_token": "__start__", "eos_token": "__end__"} lowercase : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) lowercase : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_a ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_a ) ) def __magic_name__ ( self , **_a ): kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **_a ) def __magic_name__ ( self , _a ): lowercase : Optional[Any] = "adapt act apte" lowercase : Any = "adapt act apte" return input_text, output_text def __magic_name__ ( self ): lowercase : Union[str, Any] = BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowercase : Any = "adapt act apte" lowercase : int = ["adapt", "act", "ap@@", "te"] lowercase : Optional[Any] = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) lowercase : Dict = [tokenizer.bos_token] + tokens + [tokenizer.eos_token] lowercase : List[Any] = [0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , _a ) def __magic_name__ ( self ): lowercase : List[str] = BlenderbotSmallTokenizer.from_pretrained("facebook/blenderbot-90M" ) assert tok("sam" ).input_ids == [1_384] lowercase : Any = "I am a small frog." lowercase : Tuple = tok([src_text] , padding=_a , truncation=_a )["input_ids"] lowercase : List[Any] = tok.batch_decode(_a , skip_special_tokens=_a , clean_up_tokenization_spaces=_a )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def __magic_name__ ( self ): lowercase : List[Any] = BlenderbotSmallTokenizer.from_pretrained("facebook/blenderbot-90M" ) lowercase : Union[str, Any] = "I am a small frog ." lowercase : int = "." lowercase : Optional[Any] = tok(_a )["input_ids"] lowercase : Optional[int] = tok(_a )["input_ids"] assert encoded[-1] == encoded_dot[0]
202
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch _A : List[Any] = logging.get_logger(__name__) class a__ ( a_ ): __lowerCAmelCase = ["""pixel_values"""] def __init__( self , _a = True , _a = None , _a = PILImageResampling.BILINEAR , _a = True , _a = 1 / 255 , _a = True , _a = None , _a = True , **_a , ): super().__init__(**_a ) lowercase : Optional[Any] = size if size is not None else {"shortest_edge": 224} lowercase : List[Any] = get_size_dict(_a , default_to_square=_a ) lowercase : str = crop_size if crop_size is not None else {"height": 256, "width": 256} lowercase : List[str] = get_size_dict(_a , param_name="crop_size" ) lowercase : int = do_resize lowercase : Optional[int] = size lowercase : str = resample lowercase : List[Any] = do_rescale lowercase : Union[str, Any] = rescale_factor lowercase : Optional[int] = do_center_crop lowercase : Union[str, Any] = crop_size lowercase : Optional[Any] = do_flip_channel_order def __magic_name__ ( self , _a , _a , _a = PIL.Image.BILINEAR , _a = None , **_a , ): lowercase : List[Any] = get_size_dict(_a , default_to_square=_a ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) lowercase : Union[str, Any] = get_resize_output_image_size(_a , size=size["shortest_edge"] , default_to_square=_a ) return resize(_a , size=_a , resample=_a , data_format=_a , **_a ) def __magic_name__ ( self , _a , _a , _a = None , **_a , ): lowercase : str = get_size_dict(_a ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_a , size=(size["height"], size["width"]) , data_format=_a , **_a ) def __magic_name__ ( self , _a , _a , _a = None , **_a , ): return rescale(_a , scale=_a , data_format=_a , **_a ) def __magic_name__ ( self , _a , _a = None ): return flip_channel_order(_a , data_format=_a ) def __magic_name__ ( self , _a , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = ChannelDimension.FIRST , **_a , ): lowercase : Optional[Any] = do_resize if do_resize is not None else self.do_resize lowercase : Tuple = resample if resample is not None else self.resample lowercase : List[str] = do_rescale if do_rescale is not None else self.do_rescale lowercase : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase : Optional[int] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) lowercase : str = size if size is not None else self.size lowercase : Any = get_size_dict(_a , default_to_square=_a ) lowercase : int = crop_size if crop_size is not None else self.crop_size lowercase : Any = get_size_dict(_a , param_name="crop_size" ) lowercase : int = make_list_of_images(_a ) if not valid_images(_a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) # All transformations expect numpy arrays. lowercase : Any = [to_numpy_array(_a ) for image in images] if do_resize: lowercase : Optional[int] = [self.resize(image=_a , size=_a , resample=_a ) for image in images] if do_center_crop: lowercase : str = [self.center_crop(image=_a , size=_a ) for image in images] if do_rescale: lowercase : Union[str, Any] = [self.rescale(image=_a , scale=_a ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: lowercase : int = [self.flip_channel_order(image=_a ) for image in images] lowercase : int = [to_channel_dimension_format(_a , _a ) for image in images] lowercase : Optional[Any] = {"pixel_values": images} return BatchFeature(data=_a , tensor_type=_a ) def __magic_name__ ( self , _a , _a = None ): lowercase : Optional[int] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_a ) != len(_a ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(_a ): lowercase : Tuple = target_sizes.numpy() lowercase : List[Any] = [] for idx in range(len(_a ) ): lowercase : Dict = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=_a ) lowercase : Tuple = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_a ) else: lowercase : str = logits.argmax(dim=1 ) lowercase : Dict = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
202
1
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} lowerCamelCase = { """vocab_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), }, """merges_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), }, } lowerCamelCase = { """allenai/longformer-base-4096""": 4_096, """allenai/longformer-large-4096""": 4_096, """allenai/longformer-large-4096-finetuned-triviaqa""": 4_096, """allenai/longformer-base-4096-extra.pos.embd.only""": 4_096, """allenai/longformer-large-4096-extra.pos.embd.only""": 4_096, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def a__ ( ): UpperCAmelCase_ = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) UpperCAmelCase_ = bs[:] UpperCAmelCase_ = 0 for b in range(2**8 ): if b not in bs: bs.append(lowerCAmelCase__ ) cs.append(2**8 + n ) n += 1 UpperCAmelCase_ = [chr(lowerCAmelCase__ ) for n in cs] return dict(zip(lowerCAmelCase__ , lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = set() UpperCAmelCase_ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase_ = char return pairs class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any]="replace" , _UpperCAmelCase : List[Any]="<s>" , _UpperCAmelCase : List[Any]="</s>" , _UpperCAmelCase : Optional[int]="</s>" , _UpperCAmelCase : List[Any]="<s>" , _UpperCAmelCase : str="<unk>" , _UpperCAmelCase : List[Any]="<pad>" , _UpperCAmelCase : Union[str, Any]="<mask>" , _UpperCAmelCase : Optional[int]=False , **_UpperCAmelCase : List[Any] , ) -> Dict: '''simple docstring''' UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else bos_token UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else eos_token UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else sep_token UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else cls_token UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else unk_token UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else mask_token super().__init__( errors=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , add_prefix_space=_UpperCAmelCase , **_UpperCAmelCase , ) with open(_UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ = json.load(_UpperCAmelCase ) UpperCAmelCase_ = {v: k for k, v in self.encoder.items()} UpperCAmelCase_ = errors # how to handle errors in decoding UpperCAmelCase_ = bytes_to_unicode() UpperCAmelCase_ = {v: k for k, v in self.byte_encoder.items()} with open(_UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase_ = merges_handle.read().split("\n" )[1:-1] UpperCAmelCase_ = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase_ = dict(zip(_UpperCAmelCase , range(len(_UpperCAmelCase ) ) ) ) UpperCAmelCase_ = {} UpperCAmelCase_ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase_ = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def lowercase__ ( self : Any ) -> List[str]: '''simple docstring''' return len(self.encoder ) def lowercase__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : int ) -> Union[str, Any]: '''simple docstring''' if token in self.cache: return self.cache[token] UpperCAmelCase_ = tuple(_UpperCAmelCase ) UpperCAmelCase_ = get_pairs(_UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase_ = min(_UpperCAmelCase , key=lambda _UpperCAmelCase : self.bpe_ranks.get(_UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase_ , UpperCAmelCase_ = bigram UpperCAmelCase_ = [] UpperCAmelCase_ = 0 while i < len(_UpperCAmelCase ): try: UpperCAmelCase_ = word.index(_UpperCAmelCase , _UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase_ = j if word[i] == first and i < len(_UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase_ = tuple(_UpperCAmelCase ) UpperCAmelCase_ = new_word if len(_UpperCAmelCase ) == 1: break else: UpperCAmelCase_ = get_pairs(_UpperCAmelCase ) UpperCAmelCase_ = " ".join(_UpperCAmelCase ) UpperCAmelCase_ = word return word def lowercase__ ( self : Any , _UpperCAmelCase : Optional[Any] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = [] for token in re.findall(self.pat , _UpperCAmelCase ): UpperCAmelCase_ = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_UpperCAmelCase ).split(" " ) ) return bpe_tokens def lowercase__ ( self : List[Any] , _UpperCAmelCase : int ) -> Union[str, Any]: '''simple docstring''' return self.encoder.get(_UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase__ ( self : Tuple , _UpperCAmelCase : int ) -> Optional[Any]: '''simple docstring''' return self.decoder.get(_UpperCAmelCase ) def lowercase__ ( self : Tuple , _UpperCAmelCase : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = "".join(_UpperCAmelCase ) UpperCAmelCase_ = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def lowercase__ ( self : str , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(_UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ = os.path.join( _UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase_ = os.path.join( _UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_UpperCAmelCase , ensure_ascii=_UpperCAmelCase ) + "\n" ) UpperCAmelCase_ = 0 with open(_UpperCAmelCase , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase_ = token_index writer.write(" ".join(_UpperCAmelCase ) + "\n" ) index += 1 return vocab_file, merge_file def lowercase__ ( self : Any , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] UpperCAmelCase_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowercase__ ( self : Tuple , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None , _UpperCAmelCase : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCAmelCase , token_ids_a=_UpperCAmelCase , already_has_special_tokens=_UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCAmelCase )) + [1] return [1] + ([0] * len(_UpperCAmelCase )) + [1, 1] + ([0] * len(_UpperCAmelCase )) + [1] def lowercase__ ( self : List[Any] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase__ ( self : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict=False , **_UpperCAmelCase : Optional[int] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_UpperCAmelCase ) > 0 and not text[0].isspace()): UpperCAmelCase_ = " " + text return (text, kwargs)
241
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """RWKV/rwkv-4-169m-pile""": """https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json""", """RWKV/rwkv-4-430m-pile""": """https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json""", """RWKV/rwkv-4-1b5-pile""": """https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json""", """RWKV/rwkv-4-3b-pile""": """https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json""", """RWKV/rwkv-4-7b-pile""": """https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json""", """RWKV/rwkv-4-14b-pile""": """https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json""", """RWKV/rwkv-raven-1b5""": """https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json""", """RWKV/rwkv-raven-3b""": """https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json""", """RWKV/rwkv-raven-7b""": """https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json""", """RWKV/rwkv-raven-14b""": """https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json""", } class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = '''rwkv''' UpperCamelCase = {'''max_position_embeddings''': '''context_length'''} def __init__( self : str , _UpperCAmelCase : int=50277 , _UpperCAmelCase : Optional[Any]=1024 , _UpperCAmelCase : str=4096 , _UpperCAmelCase : Any=32 , _UpperCAmelCase : Any=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Tuple=1e-5 , _UpperCAmelCase : str=0 , _UpperCAmelCase : Any=0 , _UpperCAmelCase : Union[str, Any]=6 , _UpperCAmelCase : Dict=False , _UpperCAmelCase : Any=True , **_UpperCAmelCase : Optional[Any] , ) -> str: '''simple docstring''' UpperCAmelCase_ = vocab_size UpperCAmelCase_ = context_length UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = attention_hidden_size if attention_hidden_size is not None else hidden_size UpperCAmelCase_ = intermediate_size if intermediate_size is not None else 4 * hidden_size UpperCAmelCase_ = layer_norm_epsilon UpperCAmelCase_ = rescale_every UpperCAmelCase_ = use_cache UpperCAmelCase_ = bos_token_id UpperCAmelCase_ = eos_token_id super().__init__( tie_word_embeddings=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase )
241
1
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class a ( unittest.TestCase ): def A_ ( self : Any ): snake_case_ = torch.nn.Linear(10 , 10 ) snake_case_ = torch.optim.SGD(model.parameters() , 0.1 ) snake_case_ = Accelerator() snake_case_ = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"Accelerated optimizer pickling failed with {e}" ) AcceleratorState._reset_state()
56
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : str = logging.get_logger(__name__) a : str = { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json', 'google/bigbird-roberta-large': 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json', 'google/bigbird-base-trivia-itc': 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json', # See all BigBird models at https://huggingface.co/models?filter=big_bird } class a ( _lowerCamelCase ): snake_case_ = "big_bird" def __init__( self : Union[str, Any] , lowercase_ : List[Any]=5_0358 , lowercase_ : Tuple=768 , lowercase_ : Dict=12 , lowercase_ : str=12 , lowercase_ : Tuple=3072 , lowercase_ : Any="gelu_new" , lowercase_ : Optional[Any]=0.1 , lowercase_ : List[Any]=0.1 , lowercase_ : List[Any]=4096 , lowercase_ : List[Any]=2 , lowercase_ : List[str]=0.02 , lowercase_ : Optional[int]=1e-12 , lowercase_ : Tuple=True , lowercase_ : Tuple=0 , lowercase_ : str=1 , lowercase_ : Union[str, Any]=2 , lowercase_ : Optional[Any]=66 , lowercase_ : Optional[int]="block_sparse" , lowercase_ : Any=True , lowercase_ : List[str]=False , lowercase_ : Any=64 , lowercase_ : Tuple=3 , lowercase_ : Tuple=None , **lowercase_ : Tuple , ): super().__init__( pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , sep_token_id=lowercase_ , **lowercase_ , ) snake_case_ = vocab_size snake_case_ = max_position_embeddings snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = initializer_range snake_case_ = type_vocab_size snake_case_ = layer_norm_eps snake_case_ = use_cache snake_case_ = rescale_embeddings snake_case_ = attention_type snake_case_ = use_bias snake_case_ = block_size snake_case_ = num_random_blocks snake_case_ = classifier_dropout class a ( _lowerCamelCase ): @property def A_ ( self : str ): if self.task == "multiple-choice": snake_case_ = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: snake_case_ = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
56
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ : int = { """configuration_owlvit""": [ """OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """OwlViTConfig""", """OwlViTOnnxConfig""", """OwlViTTextConfig""", """OwlViTVisionConfig""", ], """processing_owlvit""": ["""OwlViTProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : int = ["""OwlViTFeatureExtractor"""] UpperCAmelCase_ : Dict = ["""OwlViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Tuple = [ """OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """OwlViTModel""", """OwlViTPreTrainedModel""", """OwlViTTextModel""", """OwlViTVisionModel""", """OwlViTForObjectDetection""", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys UpperCAmelCase_ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
365
"""simple docstring""" from itertools import permutations def _A (__a ) -> bool: """simple docstring""" if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False SCREAMING_SNAKE_CASE_ : List[str] = [7, 11, 13, 17] for i, test in enumerate(__a ): if (num[i + 4] * 1_00 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def _A (__a = 10 ) -> int: """simple docstring""" return sum( int(''''''.join(map(__a , __a ) ) ) for num in permutations(range(__a ) ) if is_substring_divisible(__a ) ) if __name__ == "__main__": print(f'''{solution() = }''')
318
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _a : int= get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _a : Optional[int]= 250_004 _a : str= 250_020 @require_sentencepiece @require_tokenizers class UpperCamelCase ( lowercase , unittest.TestCase ): UpperCAmelCase : List[str] = MBartaaTokenizer UpperCAmelCase : str = MBartaaTokenizerFast UpperCAmelCase : int = True UpperCAmelCase : str = True def _lowercase (self : Optional[int]) -> Tuple: super().setUp() # We have a SentencePiece fixture for testing __snake_case : Union[str, Any] = MBartaaTokenizer(_A , src_lang='en_XX' , tgt_lang='ro_RO' , keep_accents=_A) tokenizer.save_pretrained(self.tmpdirname) def _lowercase (self : Dict) -> List[Any]: __snake_case : Union[str, Any] = '<s>' __snake_case : List[str] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A) , _A) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A) , _A) def _lowercase (self : List[str]) -> Union[str, Any]: __snake_case : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(_A) , 10_54) def _lowercase (self : List[str]) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 10_54) def _lowercase (self : Dict) -> List[str]: __snake_case : Any = MBartaaTokenizer(_A , src_lang='en_XX' , tgt_lang='ro_RO' , keep_accents=_A) __snake_case : str = tokenizer.tokenize('This is a test') self.assertListEqual(_A , ['▁This', '▁is', '▁a', '▁t', 'est']) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __snake_case : Union[str, Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.') self.assertListEqual( _A , [SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.'] , ) __snake_case : List[str] = tokenizer.convert_tokens_to_ids(_A) self.assertListEqual( _A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __snake_case : Any = tokenizer.convert_ids_to_tokens(_A) self.assertListEqual( _A , [SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.'] , ) @slow def _lowercase (self : Dict) -> Dict: # fmt: off __snake_case : Any = {'input_ids': [[25_00_04, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [25_00_04, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_00_04, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='facebook/mbart-large-50' , revision='d3913889c59cd5c9e456b269c376325eabad57e2' , ) def _lowercase (self : Any) -> int: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __snake_case : List[str] = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-mbart50', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})"): __snake_case : Tuple = self.rust_tokenizer_class.from_pretrained(_A , **_A) __snake_case : List[str] = self.tokenizer_class.from_pretrained(_A , **_A) __snake_case : Union[str, Any] = tempfile.mkdtemp() __snake_case : Optional[int] = tokenizer_r.save_pretrained(_A) __snake_case : Any = tokenizer_p.save_pretrained(_A) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files)) __snake_case : Tuple = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f) self.assertSequenceEqual(_A , _A) # Checks everything loads correctly in the same way __snake_case : Union[str, Any] = tokenizer_r.from_pretrained(_A) __snake_case : List[Any] = tokenizer_p.from_pretrained(_A) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_A) # Save tokenizer rust, legacy_format=True __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = tokenizer_r.save_pretrained(_A , legacy_format=_A) __snake_case : int = tokenizer_p.save_pretrained(_A) # Checks it save with the same files self.assertSequenceEqual(_A , _A) # Checks everything loads correctly in the same way __snake_case : List[Any] = tokenizer_r.from_pretrained(_A) __snake_case : Tuple = tokenizer_p.from_pretrained(_A) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A)) shutil.rmtree(_A) # Save tokenizer rust, legacy_format=False __snake_case : int = tempfile.mkdtemp() __snake_case : Tuple = tokenizer_r.save_pretrained(_A , legacy_format=_A) __snake_case : str = tokenizer_p.save_pretrained(_A) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __snake_case : List[str] = tokenizer_r.from_pretrained(_A) __snake_case : Dict = tokenizer_p.from_pretrained(_A) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A)) shutil.rmtree(_A) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): UpperCAmelCase : Dict = """facebook/mbart-large-50-one-to-many-mmt""" UpperCAmelCase : Optional[Any] = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] UpperCAmelCase : int = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] UpperCAmelCase : Dict = [EN_CODE, 8274, 12_7873, 2_5916, 7, 8622, 2071, 438, 6_7485, 53, 18_7895, 23, 5_1712, 2] @classmethod def _lowercase (cls : int) -> List[str]: __snake_case : MBartaaTokenizer = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en_XX' , tgt_lang='ro_RO') __snake_case : Union[str, Any] = 1 return cls def _lowercase (self : Optional[Any]) -> Any: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ar_AR'] , 25_00_01) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['en_EN'] , 25_00_04) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ro_RO'] , 25_00_20) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['mr_IN'] , 25_00_38) def _lowercase (self : Optional[Any]) -> int: __snake_case : List[str] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , _A) def _lowercase (self : str) -> List[Any]: self.assertIn(_A , self.tokenizer.all_special_ids) __snake_case : Optional[int] = [RO_CODE, 8_84, 90_19, 96, 9, 9_16, 8_67_92, 36, 1_87_43, 1_55_96, 5, 2] __snake_case : str = self.tokenizer.decode(_A , skip_special_tokens=_A) __snake_case : Tuple = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_A) self.assertEqual(_A , _A) self.assertNotIn(self.tokenizer.eos_token , _A) def _lowercase (self : int) -> Optional[Any]: __snake_case : int = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , _A) __snake_case : Any = 10 __snake_case : Tuple = self.tokenizer(_A , max_length=_A , truncation=_A).input_ids[0] self.assertEqual(ids[0] , _A) self.assertEqual(ids[-1] , 2) self.assertEqual(len(_A) , _A) def _lowercase (self : str) -> int: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', 'ar_AR']) , [25_00_53, 25_00_01]) def _lowercase (self : List[str]) -> Dict: __snake_case : List[Any] = tempfile.mkdtemp() __snake_case : str = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_A) __snake_case : List[Any] = MBartaaTokenizer.from_pretrained(_A) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _A) @require_torch def _lowercase (self : List[str]) -> Optional[int]: __snake_case : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_A , return_tensors='pt') __snake_case : Union[str, Any] = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def _lowercase (self : Dict) -> Dict: __snake_case : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_A , truncation=_A , max_length=len(self.expected_src_tokens) , return_tensors='pt' , ) __snake_case : str = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id) self.assertIsInstance(_A , _A) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __snake_case : List[Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _A) self.assertEqual(2 , batch.decoder_input_ids[0, 0]) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE]) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id]) def _lowercase (self : Optional[int]) -> int: __snake_case : Optional[int] = self.tokenizer(self.src_text , padding=_A , truncation=_A , max_length=3 , return_tensors='pt') __snake_case : Union[str, Any] = self.tokenizer( text_target=self.tgt_text , padding=_A , truncation=_A , max_length=10 , return_tensors='pt') __snake_case : Tuple = targets['input_ids'] __snake_case : Dict = shift_tokens_right(_A , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def _lowercase (self : Dict) -> Optional[int]: __snake_case : Any = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='ar_AR') self.assertEqual( nested_simplify(_A) , { # en_XX, A, test, EOS 'input_ids': [[25_00_04, 62, 30_34, 2]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 25_00_01, } , )
172
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : Union[str, Any]) -> Optional[Any]: __snake_case : int = 0 def _lowercase (self : Union[str, Any]) -> Optional[Any]: __snake_case : int = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32') self.assertIsInstance(_A , _A) def _lowercase (self : Any) -> Tuple: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Union[str, Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : List[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Any = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[str]) -> Optional[int]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = CLIPConfig() # Create a dummy config file with image_proceesor_type __snake_case : int = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __snake_case : List[Any] = AutoImageProcessor.from_pretrained(_A).to_dict() config_dict.pop('image_processor_type') __snake_case : Dict = CLIPImageProcessor(**_A) # save in new folder model_config.save_pretrained(_A) config.save_pretrained(_A) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) # make sure private variable is not incorrectly saved __snake_case : Optional[int] = json.loads(config.to_json_string()) self.assertTrue('_processor_class' not in dict_as_saved) self.assertIsInstance(_A , _A) def _lowercase (self : Union[str, Any]) -> str: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = Path(_A) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) __snake_case : Any = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> str: with self.assertRaisesRegex( _A , 'clip-base is not a local folder and is not a valid model identifier'): __snake_case : Tuple = AutoImageProcessor.from_pretrained('clip-base') def _lowercase (self : Optional[int]) -> int: with self.assertRaisesRegex( _A , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'): __snake_case : int = AutoImageProcessor.from_pretrained(_A , revision='aaaaaa') def _lowercase (self : Tuple) -> Optional[int]: with self.assertRaisesRegex( _A , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __snake_case : Any = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model') def _lowercase (self : Tuple) -> Tuple: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_A): __snake_case : Optional[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') # If remote code is disabled, we can't load this config. with self.assertRaises(_A): __snake_case : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) __snake_case : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Dict = AutoImageProcessor.from_pretrained(_A , trust_remote_code=_A) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor') def _lowercase (self : int) -> str: try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A): AutoImageProcessor.register(_A , _A) with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Optional[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = CustomImageProcessor.from_pretrained(_A) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowercase (self : int) -> str: class UpperCamelCase ( lowercase ): UpperCAmelCase : List[str] = True try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # If remote code is not set, the default is to use local __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote code is disabled, we load the local one. __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote is enabled, we load from the Hub __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(not hasattr(_A , 'is_local')) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
172
1
def _a ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ) -> int: '''simple docstring''' if len(SCREAMING_SNAKE_CASE__ ) != len(SCREAMING_SNAKE_CASE__ ): raise ValueError("String lengths must match!" ) SCREAMING_SNAKE_CASE__ : Dict = 0 for chara, chara in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
355
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowerCamelCase : """simple docstring""" def __init__( self : List[Any], _UpperCAmelCase : List[str], _UpperCAmelCase : Any=2, _UpperCAmelCase : List[str]=3, _UpperCAmelCase : Union[str, Any]=4, _UpperCAmelCase : Union[str, Any]=2, _UpperCAmelCase : int=7, _UpperCAmelCase : Tuple=True, _UpperCAmelCase : int=True, _UpperCAmelCase : List[str]=True, _UpperCAmelCase : List[str]=True, _UpperCAmelCase : Tuple=9_9, _UpperCAmelCase : Any=3_6, _UpperCAmelCase : List[str]=2, _UpperCAmelCase : int=4, _UpperCAmelCase : str=3_7, _UpperCAmelCase : List[str]="gelu", _UpperCAmelCase : Optional[int]=0.1, _UpperCAmelCase : str=0.1, _UpperCAmelCase : Optional[int]=5_1_2, _UpperCAmelCase : Optional[Any]=1_6, _UpperCAmelCase : int=2, _UpperCAmelCase : Tuple=0.02, _UpperCAmelCase : Optional[int]=6, _UpperCAmelCase : List[Any]=6, _UpperCAmelCase : Any=3, _UpperCAmelCase : List[str]=4, _UpperCAmelCase : Tuple=None, _UpperCAmelCase : str=1_0_0_0, ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = parent SCREAMING_SNAKE_CASE__ : Union[str, Any] = batch_size SCREAMING_SNAKE_CASE__ : int = num_channels SCREAMING_SNAKE_CASE__ : List[Any] = image_size SCREAMING_SNAKE_CASE__ : List[Any] = patch_size SCREAMING_SNAKE_CASE__ : List[Any] = is_training SCREAMING_SNAKE_CASE__ : List[Any] = use_input_mask SCREAMING_SNAKE_CASE__ : Optional[int] = use_token_type_ids SCREAMING_SNAKE_CASE__ : List[Any] = use_labels SCREAMING_SNAKE_CASE__ : int = vocab_size SCREAMING_SNAKE_CASE__ : Union[str, Any] = hidden_size SCREAMING_SNAKE_CASE__ : Tuple = num_hidden_layers SCREAMING_SNAKE_CASE__ : int = num_attention_heads SCREAMING_SNAKE_CASE__ : List[str] = intermediate_size SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_act SCREAMING_SNAKE_CASE__ : List[str] = hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Union[str, Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : Optional[int] = max_position_embeddings SCREAMING_SNAKE_CASE__ : Union[str, Any] = type_vocab_size SCREAMING_SNAKE_CASE__ : List[str] = type_sequence_label_size SCREAMING_SNAKE_CASE__ : Dict = initializer_range SCREAMING_SNAKE_CASE__ : List[str] = coordinate_size SCREAMING_SNAKE_CASE__ : Tuple = shape_size SCREAMING_SNAKE_CASE__ : Dict = num_labels SCREAMING_SNAKE_CASE__ : List[str] = num_choices SCREAMING_SNAKE_CASE__ : Optional[Any] = scope SCREAMING_SNAKE_CASE__ : Any = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) SCREAMING_SNAKE_CASE__ : str = text_seq_length SCREAMING_SNAKE_CASE__ : Dict = (image_size // patch_size) ** 2 + 1 SCREAMING_SNAKE_CASE__ : List[str] = self.text_seq_length + self.image_seq_length def A_ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = ids_tensor([self.batch_size, self.text_seq_length], self.vocab_size ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = ids_tensor([self.batch_size, self.text_seq_length, 4], self.range_bbox ) SCREAMING_SNAKE_CASE__ : List[Any] = bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: SCREAMING_SNAKE_CASE__ : Optional[Any] = bbox[i, j, 3] SCREAMING_SNAKE_CASE__ : Optional[Any] = bbox[i, j, 1] SCREAMING_SNAKE_CASE__ : Tuple = tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: SCREAMING_SNAKE_CASE__ : Tuple = bbox[i, j, 2] SCREAMING_SNAKE_CASE__ : str = bbox[i, j, 0] SCREAMING_SNAKE_CASE__ : Dict = tmp_coordinate SCREAMING_SNAKE_CASE__ : Optional[int] = tf.constant(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ : List[Any] = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ : Dict = random_attention_mask([self.batch_size, self.text_seq_length] ) SCREAMING_SNAKE_CASE__ : Tuple = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE__ : Dict = ids_tensor([self.batch_size, self.text_seq_length], self.type_vocab_size ) SCREAMING_SNAKE_CASE__ : int = None SCREAMING_SNAKE_CASE__ : Tuple = None if self.use_labels: SCREAMING_SNAKE_CASE__ : Optional[int] = ids_tensor([self.batch_size], self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ : Dict = ids_tensor([self.batch_size, self.text_seq_length], self.num_labels ) SCREAMING_SNAKE_CASE__ : List[Any] = LayoutLMvaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, coordinate_size=self.coordinate_size, shape_size=self.shape_size, input_size=self.image_size, patch_size=self.patch_size, ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def A_ ( self : Tuple, _UpperCAmelCase : str, _UpperCAmelCase : Optional[Any], _UpperCAmelCase : Any, _UpperCAmelCase : List[Any], _UpperCAmelCase : Union[str, Any], _UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = TFLayoutLMvaModel(config=_UpperCAmelCase ) # text + image SCREAMING_SNAKE_CASE__ : str = model(_UpperCAmelCase, pixel_values=_UpperCAmelCase, training=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = model( _UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, attention_mask=_UpperCAmelCase, token_type_ids=_UpperCAmelCase, training=_UpperCAmelCase, ) SCREAMING_SNAKE_CASE__ : List[Any] = model(_UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, training=_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) # text only SCREAMING_SNAKE_CASE__ : Optional[Any] = model(_UpperCAmelCase, training=_UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only SCREAMING_SNAKE_CASE__ : Union[str, Any] = model({"pixel_values": pixel_values}, training=_UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.image_seq_length, self.hidden_size) ) def A_ ( self : List[Any], _UpperCAmelCase : List[Any], _UpperCAmelCase : int, _UpperCAmelCase : Union[str, Any], _UpperCAmelCase : Dict, _UpperCAmelCase : List[str], _UpperCAmelCase : Dict, _UpperCAmelCase : List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.num_labels SCREAMING_SNAKE_CASE__ : List[str] = TFLayoutLMvaForSequenceClassification(config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = model( _UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, attention_mask=_UpperCAmelCase, token_type_ids=_UpperCAmelCase, labels=_UpperCAmelCase, training=_UpperCAmelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def A_ ( self : List[Any], _UpperCAmelCase : str, _UpperCAmelCase : int, _UpperCAmelCase : str, _UpperCAmelCase : Any, _UpperCAmelCase : List[str], _UpperCAmelCase : List[str], _UpperCAmelCase : Dict ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.num_labels SCREAMING_SNAKE_CASE__ : Dict = TFLayoutLMvaForTokenClassification(config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Tuple = model( _UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, attention_mask=_UpperCAmelCase, token_type_ids=_UpperCAmelCase, labels=_UpperCAmelCase, training=_UpperCAmelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.text_seq_length, self.num_labels) ) def A_ ( self : Dict, _UpperCAmelCase : Optional[int], _UpperCAmelCase : Optional[Any], _UpperCAmelCase : Dict, _UpperCAmelCase : Union[str, Any], _UpperCAmelCase : int, _UpperCAmelCase : Any, _UpperCAmelCase : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = 2 SCREAMING_SNAKE_CASE__ : str = TFLayoutLMvaForQuestionAnswering(config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : int = model( _UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, attention_mask=_UpperCAmelCase, token_type_ids=_UpperCAmelCase, start_positions=_UpperCAmelCase, end_positions=_UpperCAmelCase, training=_UpperCAmelCase, ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def A_ ( self : List[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.prepare_config_and_inputs() ((SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__) ,(SCREAMING_SNAKE_CASE__)) : Optional[Any] = config_and_inputs SCREAMING_SNAKE_CASE__ : List[str] = { "input_ids": input_ids, "bbox": bbox, "pixel_values": pixel_values, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class lowerCamelCase (__lowerCamelCase , __lowerCamelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) UpperCAmelCase_ = ( {"document-question-answering": TFLayoutLMvaForQuestionAnswering, "feature-extraction": TFLayoutLMvaModel} if is_tf_available() else {} ) UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = False def A_ ( self : Optional[Any], _UpperCAmelCase : Tuple, _UpperCAmelCase : Any, _UpperCAmelCase : Dict, _UpperCAmelCase : str, _UpperCAmelCase : Optional[int] ) -> List[str]: """simple docstring""" return True def A_ ( self : Optional[int], _UpperCAmelCase : Optional[int], _UpperCAmelCase : str, _UpperCAmelCase : Dict=False ) -> dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = copy.deepcopy(_UpperCAmelCase ) if model_class in get_values(_UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : Tuple = { k: tf.tile(tf.expand_dims(_UpperCAmelCase, 1 ), (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(_UpperCAmelCase, tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(_UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : List[Any] = tf.ones(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(_UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : Tuple = tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(_UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : str = tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(_UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : Tuple = tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length), dtype=tf.intaa ) return inputs_dict def A_ ( self : Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = TFLayoutLMvaModelTester(self ) SCREAMING_SNAKE_CASE__ : List[Any] = ConfigTester(self, config_class=_UpperCAmelCase, hidden_size=3_7 ) def A_ ( self : Optional[int] ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def A_ ( self : List[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : Dict = model_class(_UpperCAmelCase ) if getattr(_UpperCAmelCase, "hf_compute_loss", _UpperCAmelCase ): # The number of elements in the loss should be the same as the number of elements in the label SCREAMING_SNAKE_CASE__ : Optional[Any] = self._prepare_for_class(inputs_dict.copy(), _UpperCAmelCase, return_labels=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[int] = prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys(), reverse=_UpperCAmelCase )[0] ] SCREAMING_SNAKE_CASE__ : Dict = added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs SCREAMING_SNAKE_CASE__ : List[Any] = self._prepare_for_class(inputs_dict.copy(), _UpperCAmelCase, return_labels=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : int = prepared_for_class.pop("input_ids" ) SCREAMING_SNAKE_CASE__ : Dict = model(_UpperCAmelCase, **_UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions SCREAMING_SNAKE_CASE__ : List[Any] = self._prepare_for_class(inputs_dict.copy(), _UpperCAmelCase, return_labels=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Tuple = prepared_for_class.pop("input_ids" ) if "labels" in prepared_for_class: SCREAMING_SNAKE_CASE__ : str = prepared_for_class["labels"].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: SCREAMING_SNAKE_CASE__ : Any = -1_0_0 SCREAMING_SNAKE_CASE__ : Dict = tf.convert_to_tensor(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Tuple = model(_UpperCAmelCase, **_UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict SCREAMING_SNAKE_CASE__ : List[Any] = self._prepare_for_class(inputs_dict.copy(), _UpperCAmelCase, return_labels=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[Any] = model(_UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple SCREAMING_SNAKE_CASE__ : int = self._prepare_for_class(inputs_dict.copy(), _UpperCAmelCase, return_labels=_UpperCAmelCase ) # Get keys that were added with the _prepare_for_class function SCREAMING_SNAKE_CASE__ : List[Any] = prepared_for_class.keys() - inputs_dict.keys() SCREAMING_SNAKE_CASE__ : List[Any] = inspect.signature(model.call ).parameters SCREAMING_SNAKE_CASE__ : Dict = list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple SCREAMING_SNAKE_CASE__ : Tuple = {0: "input_ids"} for label_key in label_keys: SCREAMING_SNAKE_CASE__ : str = signature_names.index(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = label_key SCREAMING_SNAKE_CASE__ : str = sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple SCREAMING_SNAKE_CASE__ : Any = [] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: SCREAMING_SNAKE_CASE__ : int = prepared_for_class[value] SCREAMING_SNAKE_CASE__ : List[Any] = tuple(_UpperCAmelCase ) # Send to model SCREAMING_SNAKE_CASE__ : Union[str, Any] = model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def A_ ( self : Dict ) -> int: """simple docstring""" ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : List[Any] ) -> int: """simple docstring""" ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE__ : Optional[int] = type self.model_tester.create_and_check_model(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Dict ) -> Optional[Any]: """simple docstring""" ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Dict ) -> str: """simple docstring""" ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Any ) -> int: """simple docstring""" ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) @slow def A_ ( self : Optional[int] ) -> int: """simple docstring""" for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ : List[Any] = TFLayoutLMvaModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def _a ( ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf class lowerCamelCase (unittest.TestCase ): """simple docstring""" @cached_property def A_ ( self : Tuple ) -> List[Any]: """simple docstring""" return LayoutLMvaImageProcessor(apply_ocr=_UpperCAmelCase ) if is_vision_available() else None @slow def A_ ( self : Any ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = TFLayoutLMvaModel.from_pretrained("microsoft/layoutlmv3-base" ) SCREAMING_SNAKE_CASE__ : Optional[int] = self.default_image_processor SCREAMING_SNAKE_CASE__ : Dict = prepare_img() SCREAMING_SNAKE_CASE__ : List[str] = image_processor(images=_UpperCAmelCase, return_tensors="tf" ).pixel_values SCREAMING_SNAKE_CASE__ : int = tf.constant([[1, 2]] ) SCREAMING_SNAKE_CASE__ : int = tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ), axis=0 ) # forward pass SCREAMING_SNAKE_CASE__ : Dict = model(input_ids=_UpperCAmelCase, bbox=_UpperCAmelCase, pixel_values=_UpperCAmelCase, training=_UpperCAmelCase ) # verify the logits SCREAMING_SNAKE_CASE__ : Union[str, Any] = (1, 1_9_9, 7_6_8) self.assertEqual(outputs.last_hidden_state.shape, _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[int] = tf.constant( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3], _UpperCAmelCase, atol=1E-4 ) )
191
0
import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser() parser.add_argument("--model_ckpt", type=A, default="microsoft/unixcoder-base-nine" ) parser.add_argument("--num_epochs", type=A, default=5 ) parser.add_argument("--batch_size", type=A, default=6 ) parser.add_argument("--gradient_accumulation_steps", type=A, default=1 ) parser.add_argument("--freeze", type=A, default=A ) parser.add_argument("--learning_rate", type=A, default=5E-4 ) parser.add_argument("--seed", type=A, default=0 ) parser.add_argument("--lr_scheduler_type", type=A, default="cosine" ) parser.add_argument("--num_warmup_steps", type=A, default=10 ) parser.add_argument("--weight_decay", type=A, default=0.01 ) parser.add_argument("--output_dir", type=A, default="./results" ) return parser.parse_args() __lowerCAmelCase : List[str] = load('accuracy') def __magic_name__ ( A : List[Any] ): '''simple docstring''' a , a = eval_pred a = np.argmax(A, axis=1 ) return metric.compute(predictions=A, references=A ) class snake_case__ (_UpperCamelCase ): """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : Union[str, Any] ) -> None: super().__init__() a = trainer def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict , __lowerCamelCase : int , **__lowerCamelCase : int ) -> Any: if control.should_evaluate: a = deepcopy(__lowerCamelCase ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="train" ) return control_copy def __magic_name__ ( ): '''simple docstring''' a = get_args() set_seed(args.seed ) a = load_dataset("codeparrot/codecomplex", split="train" ) a = dataset.train_test_split(test_size=0.2 ) a = train_test["test"].train_test_split(test_size=0.5 ) a = DatasetDict( { "train": train_test["train"], "test": test_validation["train"], "valid": test_validation["test"], } ) print("Loading tokenizer and model" ) a = AutoTokenizer.from_pretrained(args.model_ckpt ) a = tokenizer.eos_token a = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt, num_labels=7 ) a = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): a = False a = ClassLabel(num_classes=7, names=list(set(train_test_validation["train"]["complexity"] ) ) ) def tokenize(A : str ): a = tokenizer(example["src"], truncation=A, max_length=1024 ) a = labels.straint(example["complexity"] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } a = train_test_validation.map( A, batched=A, remove_columns=train_test_validation["train"].column_names, ) a = DataCollatorWithPadding(tokenizer=A ) a = TrainingArguments( output_dir=args.output_dir, learning_rate=args.learning_rate, lr_scheduler_type=args.lr_scheduler_type, evaluation_strategy="epoch", save_strategy="epoch", logging_strategy="epoch", per_device_train_batch_size=args.batch_size, per_device_eval_batch_size=args.batch_size, num_train_epochs=args.num_epochs, gradient_accumulation_steps=args.gradient_accumulation_steps, weight_decay=0.01, metric_for_best_model="accuracy", run_name="complexity-java", report_to="wandb", ) a = Trainer( model=A, args=A, train_dataset=tokenized_datasets["train"], eval_dataset=tokenized_datasets["valid"], tokenizer=A, data_collator=A, compute_metrics=A, ) print("Training..." ) trainer.add_callback(CustomCallback(A ) ) trainer.train() if __name__ == "__main__": main()
107
from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class snake_case__ : """simple docstring""" SCREAMING_SNAKE_CASE_ : str = field( metadata={"""help""": """The output directory where the model will be written."""} , ) SCREAMING_SNAKE_CASE_ : str = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) SCREAMING_SNAKE_CASE_ : str = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def __magic_name__ ( ): '''simple docstring''' a = HfArgumentParser((ModelArguments,) ) ((a) , ) = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: a = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: a = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: a = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: a = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed a = True a = True a = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path, decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path, encoder_config=A, decoder_config=A, ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens a = decoder_config.decoder_start_token_id a = decoder_config.pad_token_id if decoder_start_token_id is None: a = decoder_config.bos_token_id if pad_token_id is None: a = decoder_config.eos_token_id # This is necessary to make Flax's generate() work a = decoder_config.eos_token_id a = decoder_start_token_id a = pad_token_id a = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) a = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) a = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
107
1
'''simple docstring''' import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a__ : '''simple docstring''' def __init__( self , lowerCamelCase_ , lowerCamelCase_=99 , lowerCamelCase_=13 , lowerCamelCase_=7 , lowerCamelCase_=9 , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_=32 , lowerCamelCase_=5 , lowerCamelCase_=4 , lowerCamelCase_=37 , lowerCamelCase_=8 , lowerCamelCase_=0.1 , lowerCamelCase_=0.002 , lowerCamelCase_=1 , lowerCamelCase_=0 , lowerCamelCase_=0 , lowerCamelCase_=None , lowerCamelCase_=None , ) -> str: lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = encoder_seq_length lowerCAmelCase__ = decoder_seq_length # For common tests lowerCAmelCase__ = self.decoder_seq_length lowerCAmelCase__ = is_training lowerCAmelCase__ = use_attention_mask lowerCAmelCase__ = use_labels lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = d_ff lowerCAmelCase__ = relative_attention_num_buckets lowerCAmelCase__ = dropout_rate lowerCAmelCase__ = initializer_factor lowerCAmelCase__ = eos_token_id lowerCAmelCase__ = pad_token_id lowerCAmelCase__ = decoder_start_token_id lowerCAmelCase__ = None lowerCAmelCase__ = decoder_layers def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: return TaConfig.from_pretrained('''google/umt5-base''' ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , ) -> List[str]: if attention_mask is None: lowerCAmelCase__ = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCAmelCase__ = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCAmelCase__ = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase_ ) if decoder_head_mask is None: lowerCAmelCase__ = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase_ ) if cross_attn_head_mask is None: lowerCAmelCase__ = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase_ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def __SCREAMING_SNAKE_CASE ( self ) -> Dict: lowerCAmelCase__ = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCAmelCase__ = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCAmelCase__ = input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ = self.get_config() lowerCAmelCase__ = config.num_attention_heads lowerCAmelCase__ = self.prepare_inputs_dict(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) return config, input_dict def __SCREAMING_SNAKE_CASE ( self ) -> List[Any]: lowerCAmelCase__ , lowerCAmelCase__ = self.prepare_config_and_inputs() return config, inputs_dict def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: return TaConfig( vocab_size=1_66 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , ) -> Dict: lowerCAmelCase__ = UMTaModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ = model( input_ids=lowerCamelCase_ , decoder_input_ids=lowerCamelCase_ , attention_mask=lowerCamelCase_ , decoder_attention_mask=lowerCamelCase_ , ) lowerCAmelCase__ = model(input_ids=lowerCamelCase_ , decoder_input_ids=lowerCamelCase_ ) lowerCAmelCase__ = result.last_hidden_state lowerCAmelCase__ = result.past_key_values lowerCAmelCase__ = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase_ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , ) -> int: lowerCAmelCase__ = UMTaModel(config=lowerCamelCase_ ).get_decoder().to(lowerCamelCase_ ).eval() # first forward pass lowerCAmelCase__ = model(lowerCamelCase_ , use_cache=lowerCamelCase_ ) lowerCAmelCase__ = model(lowerCamelCase_ ) lowerCAmelCase__ = model(lowerCamelCase_ , use_cache=lowerCamelCase_ ) self.parent.assertTrue(len(lowerCamelCase_ ) == len(lowerCamelCase_ ) ) self.parent.assertTrue(len(lowerCamelCase_ ) == len(lowerCamelCase_ ) + 1 ) lowerCAmelCase__ , lowerCAmelCase__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase__ = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCAmelCase__ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ = model(lowerCamelCase_ )['''last_hidden_state'''] lowerCAmelCase__ = model(lowerCamelCase_ , past_key_values=lowerCamelCase_ )['''last_hidden_state'''] # select random slice lowerCAmelCase__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ = output_from_no_past[:, -1, random_slice_idx].detach() lowerCAmelCase__ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase_ , lowerCamelCase_ , atol=1e-3 ) ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , ) -> List[str]: lowerCAmelCase__ = UMTaModel(config=lowerCamelCase_ ).to(lowerCamelCase_ ).half().eval() lowerCAmelCase__ = model(**lowerCamelCase_ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(lowerCamelCase_ ).any().item() ) @require_torch class a__ ( a__ , a__ , a__ , unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) lowercase__ : List[Any] = (UMTaForConditionalGeneration,) if is_torch_available() else () lowercase__ : Dict = ( { "conversational": UMTaForConditionalGeneration, "feature-extraction": UMTaModel, "summarization": UMTaForConditionalGeneration, "text2text-generation": UMTaForConditionalGeneration, "translation": UMTaForConditionalGeneration, "question-answering": UMTaForQuestionAnswering, } if is_torch_available() else {} ) lowercase__ : Optional[int] = True lowercase__ : Tuple = False lowercase__ : Optional[int] = False lowercase__ : Optional[Any] = True lowercase__ : Optional[int] = True # The small UMT5 model needs higher percentages for CPU/MP tests lowercase__ : int = [0.8, 0.9] def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: lowerCAmelCase__ = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase_ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=lowerCamelCase_ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: lowerCAmelCase__ = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ = config_and_inputs[0] lowerCAmelCase__ = UMTaForConditionalGeneration(lowerCamelCase_ ).eval() model.to(lowerCamelCase_ ) lowerCAmelCase__ = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase_ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase_ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase_ ), } for attn_name, (name, mask) in zip(lowerCamelCase_ , head_masking.items() ): lowerCAmelCase__ = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCAmelCase__ = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase_ ) lowerCAmelCase__ = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase_ , return_dict_in_generate=lowerCamelCase_ , **lowerCamelCase_ , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCAmelCase__ = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def __SCREAMING_SNAKE_CASE ( self ) -> int: pass @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): '''simple docstring''' @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def __SCREAMING_SNAKE_CASE ( self ) -> Dict: lowerCAmelCase__ = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=lowerCamelCase_ ).to(lowerCamelCase_ ) lowerCAmelCase__ = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=lowerCamelCase_ , legacy=lowerCamelCase_ ) lowerCAmelCase__ = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] lowerCAmelCase__ = tokenizer(lowerCamelCase_ , return_tensors='''pt''' , padding=lowerCamelCase_ ).input_ids # fmt: off lowerCAmelCase__ = torch.tensor( [ [ 3_85_30, 21_07_03, 25_62_99, 14_10, 25_62_98, 2_74, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 8_26, 3_21, 6_71, 2_59_22, 25_62_99, 2_74, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 14_60, 3_39, 3_12, 1_90_14, 1_06_20, 7_58, 25_62_99, 23_55,2_74, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 5_17, 25_62_99, 1_48_69, 2_81, 3_01, 25_62_98, 2_75, 11_99_83,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 3_20, 25_62_99, 1_48_69, 2_81, 22_34, 2_89, 22_75, 3_33,6_13_91, 2_89, 25_62_98, 5_43, 25_62_97, 16_87_14, 3_29, 25_62_96,2_74, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = model.generate(input_ids.to(lowerCamelCase_ ) ) lowerCAmelCase__ = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] lowerCAmelCase__ = tokenizer.batch_decode(lowerCamelCase_ ) self.assertEqual(lowerCamelCase_ , lowerCamelCase_ )
228
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __UpperCAmelCase = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' __UpperCAmelCase = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' __UpperCAmelCase = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def _snake_case ( A , A ) -> List[Any]: return float((preds == labels).mean() ) def _snake_case ( A , A , A="binary" ) -> int: lowerCAmelCase__ = simple_accuracy(A , A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=A , average=A ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( A , A ) -> List[Any]: lowerCAmelCase__ = {} for id_pred, label in zip(A , A ): lowerCAmelCase__ = F"""{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}""" lowerCAmelCase__ = id_pred['''prediction'''] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase__ = [(pred, label)] lowerCAmelCase__ , lowerCAmelCase__ = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase__ , lowerCAmelCase__ = zip(*A ) lowerCAmelCase__ = fa_score(y_true=A , y_pred=A , average='''macro''' ) fas.append(A ) lowerCAmelCase__ = int(sum(pred == label for pred, label in preds_labels ) == len(A ) ) ems.append(A ) lowerCAmelCase__ = float(sum(A ) / len(A ) ) lowerCAmelCase__ = sum(A ) / len(A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=[id_pred['''prediction'''] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self ) -> Dict: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if not self.config_name == '''record''' and not self.config_name == '''multirc''' else None , ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "prediction_text": datasets.Value('''string''' ), }, "references": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "answers": datasets.Sequence(datasets.Value('''string''' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('''int64''' ), "paragraph": datasets.Value('''int64''' ), "question": datasets.Value('''int64''' ), }, "prediction": datasets.Value('''int64''' ), }, "references": datasets.Value('''int64''' ), } else: return { "predictions": datasets.Value('''int64''' ), "references": datasets.Value('''int64''' ), } def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCamelCase_ , lowerCamelCase_ )} elif self.config_name == "cb": return acc_and_fa(lowerCamelCase_ , lowerCamelCase_ , fa_avg='''macro''' ) elif self.config_name == "record": lowerCAmelCase__ = [ { '''qas''': [ {'''id''': ref['''idx''']['''query'''], '''answers''': [{'''text''': ans} for ans in ref['''answers''']]} for ref in references ] } ] lowerCAmelCase__ = {pred['''idx''']['''query''']: pred['''prediction_text'''] for pred in predictions} return evaluate_record(lowerCamelCase_ , lowerCamelCase_ )[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCamelCase_ , lowerCamelCase_ ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCamelCase_ , lowerCamelCase_ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' )
228
1
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''' , [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 10, '''max_num_jobs''': 1}, [range(10 )]), ({'''num_shards''': 10, '''max_num_jobs''': 10}, [range(lowercase_ , i + 1 ) for i in range(10 )]), ({'''num_shards''': 1, '''max_num_jobs''': 10}, [range(1 )]), ({'''num_shards''': 10, '''max_num_jobs''': 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({'''num_shards''': 3, '''max_num_jobs''': 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> str: """simple docstring""" A__ = _distribute_shards(**lowercase_ ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''' , [ ({'''foo''': 0}, 10, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ] , ) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> str: """simple docstring""" A__ = _split_gen_kwargs(lowercase_ , lowercase_ ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''' , [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ] , ) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> Tuple: """simple docstring""" if expected is RuntimeError: with pytest.raises(lowercase_ ): _number_of_shards_in_gen_kwargs(lowercase_ ) else: A__ = _number_of_shards_in_gen_kwargs(lowercase_ ) assert out == expected
14
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _lowerCAmelCase = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' _lowerCAmelCase = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' _lowerCAmelCase = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return float((preds == labels).mean() ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase="binary" ): """simple docstring""" lowerCAmelCase__ : Any = simple_accuracy(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : Tuple = float(fa_score(y_true=UpperCamelCase , y_pred=UpperCamelCase , average=UpperCamelCase ) ) return { "accuracy": acc, "f1": fa, } def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : List[str] = {} for id_pred, label in zip(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : str = f"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" lowerCAmelCase__ : Dict = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase__ : Optional[int] = [(pred, label)] lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = zip(*UpperCamelCase ) lowerCAmelCase__ : List[Any] = fa_score(y_true=UpperCamelCase , y_pred=UpperCamelCase , average="""macro""" ) fas.append(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = int(sum(pred == label for pred, label in preds_labels ) == len(UpperCamelCase ) ) ems.append(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = float(sum(UpperCamelCase ) / len(UpperCamelCase ) ) lowerCAmelCase__ : List[Any] = sum(UpperCamelCase ) / len(UpperCamelCase ) lowerCAmelCase__ : Dict = float(fa_score(y_true=UpperCamelCase , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_( datasets.Metric ): '''simple docstring''' def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features(self._get_feature_types() ) ,codebase_urls=[] ,reference_urls=[] ,format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None ,) def UpperCAmelCase_ ( self ) -> str: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> Any: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__UpperCAmelCase ,__UpperCAmelCase )} elif self.config_name == "cb": return acc_and_fa(__UpperCAmelCase ,__UpperCAmelCase ,fa_avg="""macro""" ) elif self.config_name == "record": lowerCAmelCase__ : Optional[Any] = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] lowerCAmelCase__ : Union[str, Any] = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__UpperCAmelCase ,__UpperCAmelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__UpperCAmelCase ,__UpperCAmelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__UpperCAmelCase ,__UpperCAmelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
37
0
import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __a: """simple docstring""" def __init__( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=13 ,_SCREAMING_SNAKE_CASE=10 ,_SCREAMING_SNAKE_CASE=3 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=32 ,_SCREAMING_SNAKE_CASE=5 ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=37 ,_SCREAMING_SNAKE_CASE="gelu" ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=10 ,_SCREAMING_SNAKE_CASE=0.02 ,_SCREAMING_SNAKE_CASE=0.9 ,_SCREAMING_SNAKE_CASE=None ,) -> int: UpperCAmelCase_ : Tuple = parent UpperCAmelCase_ : Any = batch_size UpperCAmelCase_ : Any = image_size UpperCAmelCase_ : Dict = num_channels UpperCAmelCase_ : Optional[int] = patch_size UpperCAmelCase_ : int = tubelet_size UpperCAmelCase_ : Union[str, Any] = num_frames UpperCAmelCase_ : Any = is_training UpperCAmelCase_ : Optional[int] = use_labels UpperCAmelCase_ : Optional[Any] = hidden_size UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : Any = intermediate_size UpperCAmelCase_ : Union[str, Any] = hidden_act UpperCAmelCase_ : Dict = hidden_dropout_prob UpperCAmelCase_ : Any = attention_probs_dropout_prob UpperCAmelCase_ : int = type_sequence_label_size UpperCAmelCase_ : Optional[int] = initializer_range UpperCAmelCase_ : Union[str, Any] = mask_ratio UpperCAmelCase_ : Union[str, Any] = scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame UpperCAmelCase_ : List[str] = (image_size // patch_size) ** 2 UpperCAmelCase_ : List[str] = (num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos UpperCAmelCase_ : Tuple = int(mask_ratio * self.seq_length ) def a__ ( self ) -> List[Any]: UpperCAmelCase_ : Union[str, Any] = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : Union[str, Any] = None if self.use_labels: UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : int = self.get_config() return config, pixel_values, labels def a__ ( self ) -> Tuple: return VideoMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,num_frames=self.num_frames ,tubelet_size=self.tubelet_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=_SCREAMING_SNAKE_CASE ,initializer_range=self.initializer_range ,) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ : int = VideoMAEModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Any = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ : List[Any] = VideoMAEForPreTraining(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch UpperCAmelCase_ : List[str] = torch.ones((self.num_masks,) ) UpperCAmelCase_ : List[Any] = torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) UpperCAmelCase_ : List[Any] = mask.expand(self.batch_size ,-1 ).bool() UpperCAmelCase_ : Union[str, Any] = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # model only returns predictions for masked patches UpperCAmelCase_ : List[str] = mask.sum().item() UpperCAmelCase_ : Optional[int] = 3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_masked_patches, decoder_num_labels) ) def a__ ( self ) -> List[str]: UpperCAmelCase_ : Any = self.prepare_config_and_inputs() UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_ : Optional[int] = config_and_inputs UpperCAmelCase_ : Dict = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __a( _a , _a , unittest.TestCase ): """simple docstring""" lowerCAmelCase = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) lowerCAmelCase = ( {'''feature-extraction''': VideoMAEModel, '''video-classification''': VideoMAEForVideoClassification} if is_torch_available() else {} ) lowerCAmelCase = False lowerCAmelCase = False lowerCAmelCase = False lowerCAmelCase = False def a__ ( self ) -> Any: UpperCAmelCase_ : Dict = VideoMAEModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self ,config_class=_SCREAMING_SNAKE_CASE ,has_text_modality=_SCREAMING_SNAKE_CASE ,hidden_size=37 ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=False ) -> int: UpperCAmelCase_ : Union[str, Any] = copy.deepcopy(_SCREAMING_SNAKE_CASE ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch UpperCAmelCase_ : Tuple = torch.ones((self.model_tester.num_masks,) ) UpperCAmelCase_ : Union[str, Any] = torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) UpperCAmelCase_ : int = mask.expand(self.model_tester.batch_size ,-1 ).bool() UpperCAmelCase_ : Union[str, Any] = bool_masked_pos.to(_SCREAMING_SNAKE_CASE ) if return_labels: if model_class in [ *get_values(_SCREAMING_SNAKE_CASE ), ]: UpperCAmelCase_ : Any = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=_SCREAMING_SNAKE_CASE ) return inputs_dict def a__ ( self ) -> Any: self.config_tester.run_common_tests() @unittest.skip(reason='''VideoMAE does not use inputs_embeds''' ) def a__ ( self ) -> int: pass def a__ ( self ) -> List[Any]: UpperCAmelCase_, UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Any = model_class(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) UpperCAmelCase_ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE ,nn.Linear ) ) def a__ ( self ) -> Any: UpperCAmelCase_, UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Tuple = model_class(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : Any = [*signature.parameters.keys()] UpperCAmelCase_ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> Any: UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> List[str]: UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_SCREAMING_SNAKE_CASE ) @slow def a__ ( self ) -> Optional[Any]: for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : List[str] = VideoMAEModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> Any: if not self.has_attentions: pass else: UpperCAmelCase_, UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Union[str, Any] = True for model_class in self.all_model_classes: UpperCAmelCase_ : Any = self.model_tester.seq_length - self.model_tester.num_masks UpperCAmelCase_ : Tuple = ( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) UpperCAmelCase_ : Optional[int] = True UpperCAmelCase_ : Optional[int] = False UpperCAmelCase_ : int = True UpperCAmelCase_ : Optional[Any] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Tuple = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ : int = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) ,self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase_ : Optional[int] = True UpperCAmelCase_ : Any = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCAmelCase_ : List[Any] = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ : Tuple = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) UpperCAmelCase_ : Optional[Any] = len(_SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine UpperCAmelCase_ : List[str] = True UpperCAmelCase_ : int = True UpperCAmelCase_ : int = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCAmelCase_ : str = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ) self.assertEqual(out_len + 1 ,len(_SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ : List[str] = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) def a__ ( self ) -> str: def check_hidden_states_output(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : int = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Any = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ : Optional[int] = outputs.hidden_states UpperCAmelCase_ : List[Any] = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(_SCREAMING_SNAKE_CASE ) ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = self.model_tester.seq_length - self.model_tester.num_masks UpperCAmelCase_ : List[Any] = num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[seq_length, self.model_tester.hidden_size] ,) UpperCAmelCase_, UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Union[str, Any] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : Any = True check_hidden_states_output(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def a__ ( self ) -> Dict: pass def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase_ : Tuple = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) UpperCAmelCase_ : Optional[int] = np.load(_lowercase ) return list(_lowercase ) @require_torch @require_vision class __a( unittest.TestCase ): """simple docstring""" @cached_property def a__ ( self ) -> Dict: # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] ,image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def a__ ( self ) -> Any: UpperCAmelCase_ : Optional[int] = VideoMAEForVideoClassification.from_pretrained('''MCG-NJU/videomae-base-finetuned-kinetics''' ).to( _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = self.default_image_processor UpperCAmelCase_ : Optional[Any] = prepare_video() UpperCAmelCase_ : int = image_processor(_SCREAMING_SNAKE_CASE ,return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): UpperCAmelCase_ : int = model(**_SCREAMING_SNAKE_CASE ) # verify the logits UpperCAmelCase_ : List[str] = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[Any] = torch.tensor([0.36_69, -0.06_88, -0.24_21] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,_SCREAMING_SNAKE_CASE ,atol=1e-4 ) ) @slow def a__ ( self ) -> Union[str, Any]: UpperCAmelCase_ : Optional[Any] = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ).to(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = self.default_image_processor UpperCAmelCase_ : Dict = prepare_video() UpperCAmelCase_ : List[str] = image_processor(_SCREAMING_SNAKE_CASE ,return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # add boolean mask, indicating which patches to mask UpperCAmelCase_ : Dict = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' ,filename='''bool_masked_pos.pt''' ) UpperCAmelCase_ : str = torch.load(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): UpperCAmelCase_ : List[str] = model(**_SCREAMING_SNAKE_CASE ) # verify the logits UpperCAmelCase_ : str = torch.Size([1, 1_408, 1_536] ) UpperCAmelCase_ : Union[str, Any] = torch.tensor( [[0.79_94, 0.96_12, 0.85_08], [0.74_01, 0.89_58, 0.83_02], [0.58_62, 0.74_68, 0.73_25]] ,device=_SCREAMING_SNAKE_CASE ) self.assertEqual(outputs.logits.shape ,_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] ,_SCREAMING_SNAKE_CASE ,atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) UpperCAmelCase_ : Dict = torch.tensor([0.51_42] ,device=_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.loss ,_SCREAMING_SNAKE_CASE ,atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) UpperCAmelCase_ : Optional[int] = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ,norm_pix_loss=_SCREAMING_SNAKE_CASE ).to( _SCREAMING_SNAKE_CASE ) with torch.no_grad(): UpperCAmelCase_ : List[Any] = model(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = torch.tensor(torch.tensor([0.64_69] ) ,device=_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.loss ,_SCREAMING_SNAKE_CASE ,atol=1e-4 ) )
235
from dataclasses import dataclass, field from typing import Optional @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be trained.'''} ) lowerCAmelCase = field( default='''./''' , metadata={'''help''': '''Save dir where model repo is cloned and models updates are saved to.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot-clean-train''' , metadata={'''help''': '''Name or path of training dataset.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot-clean-valid''' , metadata={'''help''': '''Name or path of validation dataset.'''} ) lowerCAmelCase = field(default=2 , metadata={'''help''': '''Batch size for training.'''} ) lowerCAmelCase = field(default=2 , metadata={'''help''': '''Batch size for evaluation.'''} ) lowerCAmelCase = field(default=0.1 , metadata={'''help''': '''Value of weight decay.'''} ) lowerCAmelCase = field( default=1_0000 , metadata={'''help''': '''Size of buffer used to shuffle streaming dataset.'''} ) lowerCAmelCase = field(default=2E-4 , metadata={'''help''': '''Learning rate fo training.'''} ) lowerCAmelCase = field(default='''cosine''' , metadata={'''help''': '''Learning rate.'''} ) lowerCAmelCase = field( default=750 , metadata={'''help''': '''Number of warmup steps in the learning rate schedule.'''} ) lowerCAmelCase = field( default=16 , metadata={'''help''': '''Number of gradient accumulation steps.'''} ) lowerCAmelCase = field( default=_a , metadata={'''help''': '''Use gradient checkpointing to reduce memory footprint.'''} ) lowerCAmelCase = field(default=5_0000 , metadata={'''help''': '''Maximum number of training steps.'''} ) lowerCAmelCase = field( default=-1 , metadata={'''help''': '''Maximum number of evaluation steps. If -1 the full dataset is evaluated.'''} ) lowerCAmelCase = field(default=1024 , metadata={'''help''': '''Sequence lengths used for training.'''} ) lowerCAmelCase = field(default=1 , metadata={'''help''': '''Training seed.'''} ) lowerCAmelCase = field( default=1024 , metadata={'''help''': '''Interval to save checkpoints. Measured as number of forward passes not training steps.'''} , ) lowerCAmelCase = field( default=_a , metadata={'''help''': '''States path if the training should continue from a checkpoint folder.'''} ) lowerCAmelCase = field(default=_a , metadata={'''help''': '''If True the data is pretokenized.'''} ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be evaluated.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot-clean-valid''' , metadata={'''help''': '''Name or path of validation dataset.'''} ) lowerCAmelCase = field(default=2 , metadata={'''help''': '''Batch size used for evaluation.'''} ) lowerCAmelCase = field( default=-1 , metadata={'''help''': '''Maximum number of evaluation steps. If -1 the full dataset is evaluated.'''} ) lowerCAmelCase = field(default=1024 , metadata={'''help''': '''Length of sequences to be evaluated.'''} ) lowerCAmelCase = field(default=1 , metadata={'''help''': '''Random seed used for evaluation.'''} ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be evaluated.'''} ) lowerCAmelCase = field(default=_a , metadata={'''help''': '''Number of workers used for code evaluation.'''} ) lowerCAmelCase = field( default=_a , metadata={'''help''': '''The number of human-eval tasks to run. If not included all tasks are evaluated.'''} , ) lowerCAmelCase = field( default=_a , metadata={'''help''': '''Sample from the language model\'s output distribution.'''} ) lowerCAmelCase = field(default=0.2 , metadata={'''help''': '''Sampling temperature used for generation.'''} ) lowerCAmelCase = field(default=256 , metadata={'''help''': '''Maximum number of newly generated tokens.'''} ) lowerCAmelCase = field(default=0 , metadata={'''help''': '''Top-k parameter used for generation.'''} ) lowerCAmelCase = field(default=0.95 , metadata={'''help''': '''Top-p parameter used for nucleus sampling.'''} ) lowerCAmelCase = field(default=10 , metadata={'''help''': '''Number of generations to run in parallel.'''} ) lowerCAmelCase = field( default=200 , metadata={'''help''': '''Number of completions to generate for each sample.'''} ) lowerCAmelCase = field(default=1 , metadata={'''help''': '''Random seed used for evaluation.'''} ) lowerCAmelCase = field( default='''eval_results.json''' , metadata={'''help''': '''Random seed used for evaluation.'''} ) lowerCAmelCase = field( default='''0''' , metadata={'''help''': '''Allow `code_eval` to execute Python code on machine'''} ) lowerCAmelCase = field( default=-1 , metadata={ '''help''': ( '''Determine which device to run the `text-generation` Pipeline on. -1 is CPU and any zero or positive''' ''' number corresponds to which GPU device id to run on.''' ) } , ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default=_a , metadata={ '''help''': '''The number of CPU cores to use for parallel preprocessing. Default uses the maximum available.''' } , ) lowerCAmelCase = field( default='''transformersbook/codeparrot''' , metadata={'''help''': '''Folder or name of dataset to process.'''} ) lowerCAmelCase = field( default='''codeparrot-clean''' , metadata={'''help''': '''Folder to save processed processed dataset.'''} ) lowerCAmelCase = field( default=10_0000 , metadata={'''help''': '''Number of files to save per JSON output file.'''} ) lowerCAmelCase = field(default='''content''' , metadata={'''help''': '''Column containing text data to process.'''} ) lowerCAmelCase = field( default=1000 , metadata={'''help''': '''Maximum line length in file, otherwise file is filtered.'''} ) lowerCAmelCase = field( default=100 , metadata={'''help''': '''Maximum mean line length in file, otherwise file is filtered.'''} ) lowerCAmelCase = field( default=0.25 , metadata={'''help''': '''Maximum fraction of non-alphanumeric characters, otherwise file is filtered.'''} ) lowerCAmelCase = field( default=1.5 , metadata={'''help''': '''Minimum character token ratio for the file, otherwise file is filtered.'''} ) lowerCAmelCase = field( default=0.7 , metadata={'''help''': '''Probability for filtering config, test and uncommon files.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Name or path to the tokenizer.'''} , ) lowerCAmelCase = field( default=_a , metadata={'''help''': '''If True, near-duplicate samples are removed.'''} ) lowerCAmelCase = field( default=0.85 , metadata={'''help''': '''Jaccard threshold for near-duplicate samples.'''} ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''gpt2''' , metadata={'''help''': '''Base tokenizer to build new tokenizer from.'''} ) lowerCAmelCase = field( default='''transformersbook/codeparrot-train''' , metadata={'''help''': '''Dataset to train tokenizer on.'''} ) lowerCAmelCase = field(default='''content''' , metadata={'''help''': '''Column containing text data to process.'''} ) lowerCAmelCase = field(default=20_0000 , metadata={'''help''': '''Number of examples to train tokenizer on.'''} ) lowerCAmelCase = field( default=3_2768 , metadata={'''help''': '''Number of examples to train the tokenizer on.'''} ) lowerCAmelCase = field(default='''codeparrot''' , metadata={'''help''': '''Name of new tokenizer.'''} ) lowerCAmelCase = field(default=_a , metadata={'''help''': '''Push saved tokenizer to the hub.'''} ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Name or path to the tokenizer.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot-clean-train''' , metadata={'''help''': '''Name or path to the dataset to pretokenize.'''} ) lowerCAmelCase = field( default='''tokenized-codeparrot-train''' , metadata={'''help''': '''Repo name of the pretokenized data.'''} ) lowerCAmelCase = field(default=_a , metadata={'''help''': '''Number of workers used for code evaluation.'''} ) @dataclass class __a: """simple docstring""" lowerCAmelCase = field( default='''gpt2-large''' , metadata={'''help''': '''Configuration to use for model initialization.'''} ) lowerCAmelCase = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Tokenizer attached to model.'''} ) lowerCAmelCase = field(default='''codeparrot''' , metadata={'''help''': '''Name of the created model.'''} ) lowerCAmelCase = field(default=_a , metadata={'''help''': '''Push saved tokenizer to the hub.'''} )
235
1