code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
def UpperCAmelCase_ ( __lowerCAmelCase ) -> Optional[Any]: stooge(_SCREAMING_SNAKE_CASE , 0 , len(_SCREAMING_SNAKE_CASE ) - 1 ) return arr def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: __lowercase , __lowercase : Any = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: __lowercase : Any = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (h - t) ) # Recursively sort last 2/3 elements stooge(_SCREAMING_SNAKE_CASE , i + t , (_SCREAMING_SNAKE_CASE) ) # Recursively sort first 2/3 elements stooge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (h - t) ) if __name__ == "__main__": __lowerCAmelCase : Optional[Any] = input("Enter numbers separated by a comma:\n").strip() __lowerCAmelCase : Optional[Any] = [int(item) for item in user_input.split(",")] print(stooge_sort(unsorted))
156
from argparse import ArgumentParser from .env import EnvironmentCommand def __lowercase ( ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) SCREAMING_SNAKE_CASE = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(_SCREAMING_SNAKE_CASE ) # Let's go SCREAMING_SNAKE_CASE = parser.parse_args() if not hasattr(_SCREAMING_SNAKE_CASE , """func""" ): parser.print_help() exit(1 ) # Run SCREAMING_SNAKE_CASE = args.func(_SCREAMING_SNAKE_CASE ) service.run() if __name__ == "__main__": main()
296
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
48
'''simple docstring''' from sklearn.metrics import fa_score import datasets lowerCamelCase = """ The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) """ lowerCamelCase = """ Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `'binary'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`. - 'binary': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - 'micro': Calculate metrics globally by counting the total true positives, false negatives and false positives. - 'macro': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - 'weighted': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - 'samples': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {'f1': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results['f1'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results['f1'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"macro\") >>> print(round(results['f1'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"micro\") >>> print(round(results['f1'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"weighted\") >>> print(round(results['f1'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {'f1': array([0.8, 0. , 0. ])} """ lowerCamelCase = """ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): '''simple docstring''' def __lowerCamelCase ( self : Any): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('int32')), 'references': datasets.Sequence(datasets.Value('int32')), } if self.config_name == 'multilabel' else { 'predictions': datasets.Value('int32'), 'references': datasets.Value('int32'), }) , reference_urls=['https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'] , ) def __lowerCamelCase ( self : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple=None , _lowerCAmelCase : Optional[int]=1 , _lowerCAmelCase : List[Any]="binary" , _lowerCAmelCase : Tuple=None): '''simple docstring''' __lowercase =fa_score( _lowerCAmelCase , _lowerCAmelCase , labels=_lowerCAmelCase , pos_label=_lowerCAmelCase , average=_lowerCAmelCase , sample_weight=_lowerCAmelCase) return {"f1": float(_lowerCAmelCase) if score.size == 1 else score}
48
1
"""simple docstring""" from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _UpperCAmelCase : '''simple docstring''' pass
102
"""simple docstring""" import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def snake_case_ ( A_ : Any ): '''simple docstring''' _lowerCamelCase : Any = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', '''decoder.output_projection.weight''', ] for k in ignore_keys: state_dict.pop(A_, A_ ) def snake_case_ ( A_ : Union[str, Any] ): '''simple docstring''' _lowerCamelCase , _lowerCamelCase : Tuple = emb.weight.shape _lowerCamelCase : Dict = nn.Linear(A_, A_, bias=A_ ) _lowerCamelCase : str = emb.weight.data return lin_layer def snake_case_ ( A_ : str, A_ : Optional[int]="facebook/mbart-large-en-ro", A_ : Union[str, Any]=False, A_ : List[str]=False ): '''simple docstring''' _lowerCamelCase : Tuple = torch.load(A_, map_location='''cpu''' )['''model'''] remove_ignore_keys_(A_ ) _lowerCamelCase : int = state_dict['''encoder.embed_tokens.weight'''].shape[0] _lowerCamelCase : Any = MBartConfig.from_pretrained(A_, vocab_size=A_ ) if mbart_aa and finetuned: _lowerCamelCase : Any = '''relu''' _lowerCamelCase : Optional[int] = state_dict['''decoder.embed_tokens.weight'''] _lowerCamelCase : Any = MBartForConditionalGeneration(A_ ) model.model.load_state_dict(A_ ) if finetuned: _lowerCamelCase : str = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
72
0
from typing import Any class lowerCamelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : Any ) -> Any: A : Union[str, Any] = data A : Any = None def __repr__( self : int ) -> str: return F"""Node({self.data})""" class lowerCamelCase_ : '''simple docstring''' def __init__( self : str ) -> int: A : int = None def __iter__( self : Optional[Any] ) -> Any: A : Tuple = self.head while node: yield node.data A : Dict = node.next def __len__( self : int ) -> int: return sum(1 for _ in self ) def __repr__( self : List[Any] ) -> str: return "->".join([str(__lowerCamelCase ) for item in self] ) def __getitem__( self : Union[str, Any] , __lowerCamelCase : int ) -> Any: if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Any , __lowerCamelCase : int , __lowerCamelCase : Any ) -> None: if not 0 <= index < len(self ): raise ValueError("list index out of range." ) A : int = self.head for _ in range(__lowerCamelCase ): A : Union[str, Any] = current.next A : Any = data def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , __lowerCamelCase : Any ) -> None: self.insert_nth(len(self ) , __lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Dict , __lowerCamelCase : Any ) -> None: self.insert_nth(0 , __lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any ) -> None: if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) A : Union[str, Any] = Node(__lowerCamelCase ) if self.head is None: A : Union[str, Any] = new_node elif index == 0: A : int = self.head # link new_node to head A : List[str] = new_node else: A : List[str] = self.head for _ in range(index - 1 ): A : Union[str, Any] = temp.next A : Dict = temp.next A : Optional[Any] = new_node def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> None: # print every node data print(self ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: return self.delete_nth(0 ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: # delete from tail return self.delete_nth(len(self ) - 1 ) def SCREAMING_SNAKE_CASE__ ( self : Any , __lowerCamelCase : int = 0 ) -> Any: if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) A : str = self.head # default first node if index == 0: A : List[Any] = self.head.next else: A : Dict = self.head for _ in range(index - 1 ): A : Tuple = temp.next A : List[Any] = temp.next A : Dict = temp.next.next return delete_node.data def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> bool: return self.head is None def SCREAMING_SNAKE_CASE__ ( self : str ) -> None: A : str = None A : int = self.head while current: # Store the current node's next node. A : Tuple = current.next # Make the current node's next point backwards A : Tuple = prev # Make the previous node be the current node A : List[Any] = current # Make the current node the next node (to progress iteration) A : Dict = next_node # Return prev in order to put the head at the end A : Union[str, Any] = prev def UpperCAmelCase ( ): A : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_lowerCamelCase ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(_lowerCamelCase ) == i linked_list.insert_nth(_lowerCamelCase , i + 1 ) assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(_lowerCamelCase ) == 9 assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): A : Optional[int] = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(-8 , 1 ) ) def UpperCAmelCase ( ): A : List[str] = [ -9, 100, Node(7734_5112 ), 'dlrow olleH', 7, 5555, 0, -192.5_5555, 'Hello, world!', 77.9, Node(10 ), None, None, 12.20, ] A : Optional[int] = LinkedList() for i in test_input: linked_list.insert_tail(_lowerCamelCase ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_lowerCamelCase ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head A : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail A : Tuple = linked_list.delete_tail() assert result == 12.2 assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list A : Optional[Any] = linked_list.delete_nth(10 ) assert result is None assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!" ) ) assert ( str(_lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_lowerCamelCase ) assert ( str(_lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_lowerCamelCase ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def UpperCAmelCase ( ): from doctest import testmod testmod() A : Dict = LinkedList() linked_list.insert_head(input("Inserting 1st at head " ).strip() ) linked_list.insert_head(input("Inserting 2nd at head " ).strip() ) print("\nPrint list:" ) linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail " ).strip() ) linked_list.insert_tail(input("Inserting 2nd at tail " ).strip() ) print("\nPrint list:" ) linked_list.print_list() print("\nDelete head" ) linked_list.delete_head() print("Delete tail" ) linked_list.delete_tail() print("\nPrint list:" ) linked_list.print_list() print("\nReverse linked list" ) linked_list.reverse() print("\nPrint list:" ) linked_list.print_list() print("\nString representation of linked list:" ) print(_lowerCamelCase ) print("\nReading/changing Node data using indexing:" ) print(f"""Element at Position 1: {linked_list[1]}""" ) A : Optional[Any] = input("Enter New Value: " ).strip() print("New list:" ) print(_lowerCamelCase ) print(f"""length of linked_list is : {len(_lowerCamelCase )}""" ) if __name__ == "__main__": main()
356
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
256
0
import random class A__ : """simple docstring""" @staticmethod def __lowercase ( lowercase) -> tuple[list[int], list[int]]: '''simple docstring''' a__ : Union[str, Any] = [ord(lowercase) for i in text] a__ : int = [] a__ : List[str] = [] for i in plain: a__ : Optional[Any] = random.randint(1 , 300) a__ : Optional[Any] = (i + k) * k cipher.append(lowercase) key.append(lowercase) return cipher, key @staticmethod def __lowercase ( lowercase , lowercase) -> str: '''simple docstring''' a__ : str = [] for i in range(len(lowercase)): a__ : Dict = int((cipher[i] - (key[i]) ** 2) / key[i]) plain.append(chr(lowercase)) return "".join(lowercase) if __name__ == "__main__": lowercase , lowercase : Optional[Any] = Onepad().encrypt("""Hello""") print(c, k) print(Onepad().decrypt(c, k))
99
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ : List[Any] = { '''configuration_distilbert''': [ '''DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''DistilBertConfig''', '''DistilBertOnnxConfig''', ], '''tokenization_distilbert''': ['''DistilBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Any = ['''DistilBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : int = [ '''DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''DistilBertForMaskedLM''', '''DistilBertForMultipleChoice''', '''DistilBertForQuestionAnswering''', '''DistilBertForSequenceClassification''', '''DistilBertForTokenClassification''', '''DistilBertModel''', '''DistilBertPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Dict = [ '''TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFDistilBertForMaskedLM''', '''TFDistilBertForMultipleChoice''', '''TFDistilBertForQuestionAnswering''', '''TFDistilBertForSequenceClassification''', '''TFDistilBertForTokenClassification''', '''TFDistilBertMainLayer''', '''TFDistilBertModel''', '''TFDistilBertPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Tuple = [ '''FlaxDistilBertForMaskedLM''', '''FlaxDistilBertForMultipleChoice''', '''FlaxDistilBertForQuestionAnswering''', '''FlaxDistilBertForSequenceClassification''', '''FlaxDistilBertForTokenClassification''', '''FlaxDistilBertModel''', '''FlaxDistilBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys lowercase__ : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
264
0
'''simple docstring''' from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None ): if version.parse(hfh.__version__ ).release < version.parse("""0.11.0""" ).release: # old versions of hfh don't url-encode the file path UpperCAmelCase__ : str = quote(__UpperCAmelCase ) return hfh.hf_hub_url(__UpperCAmelCase , __UpperCAmelCase , repo_type="""dataset""" , revision=__UpperCAmelCase )
352
'''simple docstring''' import functools def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): # Validation if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or not all(isinstance(UpperCamelCase__ , UpperCamelCase__ ) for day in days ): raise ValueError("""The parameter days should be a list of integers""" ) if len(UpperCamelCase__ ) != 3 or not all(isinstance(UpperCamelCase__ , UpperCamelCase__ ) for cost in costs ): raise ValueError("""The parameter costs should be a list of three integers""" ) if len(UpperCamelCase__ ) == 0: return 0 if min(UpperCamelCase__ ) <= 0: raise ValueError("""All days elements should be greater than 0""" ) if max(UpperCamelCase__ ) >= 3_6_6: raise ValueError("""All days elements should be less than 366""" ) UpperCAmelCase__ : Union[str, Any] = set(UpperCamelCase__ ) @functools.cache def dynamic_programming(UpperCamelCase__ ) -> int: if index > 3_6_5: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 3_0 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
283
0
'''simple docstring''' def _lowerCamelCase ( lowercase : float , lowercase : int ) -> float: if digit_amount > 0: return round(number - int(lowercase ) , lowercase ) return number - int(lowercase ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
63
'''simple docstring''' from ....utils import logging lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" def __init__( self : Tuple , __a : int , __a : Any=None , __a : Optional[int]=20_48 ): _a = config.__dict__ _a = modal_hidden_size if num_labels: _a = num_labels
63
1
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers _snake_case = 'python tqdm regex requests packaging filelock numpy tokenizers'.split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append('''dataclasses''') if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append('''importlib_metadata''') for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F'can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py') def _UpperCamelCase ( snake_case__, snake_case__=None ) -> List[Any]: require_version(deps[pkg], snake_case_ )
370
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { '''weiweishi/roc-bert-base-zh''': '''https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json''', } class _snake_case ( _lowercase ): lowerCamelCase__: Dict = "roc_bert" def __init__( self: int , __lowerCamelCase: Union[str, Any]=3_05_22 , __lowerCamelCase: int=7_68 , __lowerCamelCase: Any=12 , __lowerCamelCase: int=12 , __lowerCamelCase: Union[str, Any]=30_72 , __lowerCamelCase: Union[str, Any]="gelu" , __lowerCamelCase: Optional[int]=0.1 , __lowerCamelCase: str=0.1 , __lowerCamelCase: Any=5_12 , __lowerCamelCase: Union[str, Any]=2 , __lowerCamelCase: str=0.02 , __lowerCamelCase: int=1e-12 , __lowerCamelCase: str=True , __lowerCamelCase: int=0 , __lowerCamelCase: List[str]="absolute" , __lowerCamelCase: List[Any]=None , __lowerCamelCase: Optional[int]=True , __lowerCamelCase: List[str]=True , __lowerCamelCase: Dict=7_68 , __lowerCamelCase: Optional[int]=9_10 , __lowerCamelCase: Union[str, Any]=5_12 , __lowerCamelCase: int=2_48_58 , __lowerCamelCase: Optional[int]=True , **__lowerCamelCase: Any , ) -> List[Any]: __UpperCAmelCase : str = vocab_size __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : Optional[Any] = hidden_size __UpperCAmelCase : Optional[int] = num_hidden_layers __UpperCAmelCase : Union[str, Any] = num_attention_heads __UpperCAmelCase : List[str] = intermediate_size __UpperCAmelCase : List[Any] = hidden_act __UpperCAmelCase : List[str] = hidden_dropout_prob __UpperCAmelCase : Optional[int] = attention_probs_dropout_prob __UpperCAmelCase : Union[str, Any] = initializer_range __UpperCAmelCase : Optional[Any] = type_vocab_size __UpperCAmelCase : List[Any] = layer_norm_eps __UpperCAmelCase : Optional[int] = use_cache __UpperCAmelCase : Optional[Any] = enable_pronunciation __UpperCAmelCase : Any = enable_shape __UpperCAmelCase : Union[str, Any] = pronunciation_embed_dim __UpperCAmelCase : Optional[Any] = pronunciation_vocab_size __UpperCAmelCase : Optional[Any] = shape_embed_dim __UpperCAmelCase : List[Any] = shape_vocab_size __UpperCAmelCase : int = concat_input __UpperCAmelCase : int = position_embedding_type __UpperCAmelCase : Optional[int] = classifier_dropout super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase )
342
0
UpperCAmelCase_ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = input("""Enter message: """ ) __lowerCamelCase = input("""Enter key [alphanumeric]: """ ) __lowerCamelCase = input("""Encrypt/Decrypt [e/d]: """ ) if mode.lower().startswith("""e""" ): __lowerCamelCase = """encrypt""" __lowerCamelCase = encrypt_message(A__ , A__ ) elif mode.lower().startswith("""d""" ): __lowerCamelCase = """decrypt""" __lowerCamelCase = decrypt_message(A__ , A__ ) print(f'\n{mode.title()}ed message:' ) print(A__ ) def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' return translate_message(A__ , A__ , """encrypt""" ) def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' return translate_message(A__ , A__ , """decrypt""" ) def lowerCamelCase__ ( A__ : str , A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 0 __lowerCamelCase = key.upper() for symbol in message: __lowerCamelCase = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(A__ ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(A__ ): __lowerCamelCase = 0 else: translated.append(A__ ) return "".join(A__ ) if __name__ == "__main__": main()
12
import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'tensor(bool)': np.bool_, 'tensor(int8)': np.inta, 'tensor(uint8)': np.uinta, 'tensor(int16)': np.intaa, 'tensor(uint16)': np.uintaa, 'tensor(int32)': np.intaa, 'tensor(uint32)': np.uintaa, 'tensor(int64)': np.intaa, 'tensor(uint64)': np.uintaa, 'tensor(float16)': np.floataa, 'tensor(float)': np.floataa, 'tensor(double)': np.floataa, } class lowerCamelCase__: def __init__( self: str , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: str ): logger.info("""`diffusers.OnnxRuntimeModel` is experimental and might change in the future.""" ) __lowerCamelCase = model __lowerCamelCase = kwargs.get("""model_save_dir""" , UpperCamelCase_ ) __lowerCamelCase = kwargs.get("""latest_model_name""" , UpperCamelCase_ ) def __call__( self: Dict , **UpperCamelCase_: Any ): __lowerCamelCase = {k: np.array(UpperCamelCase_ ) for k, v in kwargs.items()} return self.model.run(UpperCamelCase_ , UpperCamelCase_ ) @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: Union[str, Path] , UpperCamelCase_: Tuple=None , UpperCamelCase_: Tuple=None ): if provider is None: logger.info("""No onnxruntime provider specified, using CPUExecutionProvider""" ) __lowerCamelCase = """CPUExecutionProvider""" return ort.InferenceSession(UpperCamelCase_ , providers=[provider] , sess_options=UpperCamelCase_ ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: Union[str, Path] , UpperCamelCase_: Optional[str] = None , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = file_name if file_name is not None else ONNX_WEIGHTS_NAME __lowerCamelCase = self.model_save_dir.joinpath(self.latest_model_name ) __lowerCamelCase = Path(UpperCamelCase_ ).joinpath(UpperCamelCase_ ) try: shutil.copyfile(UpperCamelCase_ , UpperCamelCase_ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) __lowerCamelCase = self.model_save_dir.joinpath(UpperCamelCase_ ) if src_path.exists(): __lowerCamelCase = Path(UpperCamelCase_ ).joinpath(UpperCamelCase_ ) try: shutil.copyfile(UpperCamelCase_ , UpperCamelCase_ ) except shutil.SameFileError: pass def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Union[str, os.PathLike] , **UpperCamelCase_: Optional[Any] , ): if os.path.isfile(UpperCamelCase_ ): logger.error(F'Provided path ({save_directory}) should be a directory, not a file' ) return os.makedirs(UpperCamelCase_ , exist_ok=UpperCamelCase_ ) # saving model weights/files self._save_pretrained(UpperCamelCase_ , **UpperCamelCase_ ) @classmethod def lowerCAmelCase__ ( cls: str , UpperCamelCase_: Union[str, Path] , UpperCamelCase_: Optional[Union[bool, str, None]] = None , UpperCamelCase_: Optional[Union[str, None]] = None , UpperCamelCase_: bool = False , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: Optional["ort.SessionOptions"] = None , **UpperCamelCase_: int , ): __lowerCamelCase = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(UpperCamelCase_ ): __lowerCamelCase = OnnxRuntimeModel.load_model( os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , provider=UpperCamelCase_ , sess_options=UpperCamelCase_ ) __lowerCamelCase = Path(UpperCamelCase_ ) # load model from hub else: # download model __lowerCamelCase = hf_hub_download( repo_id=UpperCamelCase_ , filename=UpperCamelCase_ , use_auth_token=UpperCamelCase_ , revision=UpperCamelCase_ , cache_dir=UpperCamelCase_ , force_download=UpperCamelCase_ , ) __lowerCamelCase = Path(UpperCamelCase_ ).parent __lowerCamelCase = Path(UpperCamelCase_ ).name __lowerCamelCase = OnnxRuntimeModel.load_model(UpperCamelCase_ , provider=UpperCamelCase_ , sess_options=UpperCamelCase_ ) return cls(model=UpperCamelCase_ , **UpperCamelCase_ ) @classmethod def lowerCAmelCase__ ( cls: Optional[int] , UpperCamelCase_: Union[str, Path] , UpperCamelCase_: bool = True , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: Optional[str] = None , **UpperCamelCase_: int , ): __lowerCamelCase = None if len(str(UpperCamelCase_ ).split("""@""" ) ) == 2: __lowerCamelCase, __lowerCamelCase = model_id.split("""@""" ) return cls._from_pretrained( model_id=UpperCamelCase_ , revision=UpperCamelCase_ , cache_dir=UpperCamelCase_ , force_download=UpperCamelCase_ , use_auth_token=UpperCamelCase_ , **UpperCamelCase_ , )
12
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer __SCREAMING_SNAKE_CASE : List[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Optional[int] = { '''vocab_file''': { '''google/electra-small-generator''': ( '''https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt''' ), '''google/electra-base-generator''': '''https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt''', '''google/electra-large-generator''': ( '''https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt''' ), '''google/electra-small-discriminator''': ( '''https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt''' ), '''google/electra-base-discriminator''': ( '''https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt''' ), '''google/electra-large-discriminator''': ( '''https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''google/electra-small-generator''': ( '''https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json''' ), '''google/electra-base-generator''': ( '''https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json''' ), '''google/electra-large-generator''': ( '''https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json''' ), '''google/electra-small-discriminator''': ( '''https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json''' ), '''google/electra-base-discriminator''': ( '''https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json''' ), '''google/electra-large-discriminator''': ( '''https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : Tuple = { '''google/electra-small-generator''': 5_1_2, '''google/electra-base-generator''': 5_1_2, '''google/electra-large-generator''': 5_1_2, '''google/electra-small-discriminator''': 5_1_2, '''google/electra-base-discriminator''': 5_1_2, '''google/electra-large-discriminator''': 5_1_2, } __SCREAMING_SNAKE_CASE : Dict = { '''google/electra-small-generator''': {'''do_lower_case''': True}, '''google/electra-base-generator''': {'''do_lower_case''': True}, '''google/electra-large-generator''': {'''do_lower_case''': True}, '''google/electra-small-discriminator''': {'''do_lower_case''': True}, '''google/electra-base-discriminator''': {'''do_lower_case''': True}, '''google/electra-large-discriminator''': {'''do_lower_case''': True}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Optional[Any] = VOCAB_FILES_NAMES lowercase__ : str = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[int] = PRETRAINED_INIT_CONFIGURATION lowercase__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Dict = ElectraTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
73
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : bool = False ) -> list[float]: if radian_mode: return [magnitude * cos(lowercase_ ), magnitude * sin(lowercase_ )] return [magnitude * cos(radians(lowercase_ ) ), magnitude * sin(radians(lowercase_ ) )] def lowerCAmelCase_( lowercase_ : NDArray[floataa] , lowercase_ : NDArray[floataa] , lowercase_ : float = 10**-1 ) -> bool: _lowerCamelCase = cross(lowercase_ , lowercase_ ) _lowerCamelCase = sum(lowercase_ ) return abs(lowercase_ ) < eps if __name__ == "__main__": # Test to check if it works __SCREAMING_SNAKE_CASE : Union[str, Any] = array( [ polar_force(718.4, 1_8_0 - 3_0), polar_force(879.54, 4_5), polar_force(1_0_0, -9_0), ] ) __SCREAMING_SNAKE_CASE : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg __SCREAMING_SNAKE_CASE : Optional[int] = array( [ polar_force(3_0 * 9.81, 1_5), polar_force(2_1_5, 1_8_0 - 4_5), polar_force(2_6_4, 9_0 - 3_0), ] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg __SCREAMING_SNAKE_CASE : str = array([[0, -2_0_0_0], [0, -1_2_0_0], [0, 1_5_6_0_0], [0, -1_2_4_0_0]]) __SCREAMING_SNAKE_CASE : str = array([[0, 0], [6, 0], [1_0, 0], [1_2, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
73
1
from __future__ import annotations from collections.abc import Iterator class SCREAMING_SNAKE_CASE__ : def __init__( self , a): lowercase__ : Dict = value lowercase__ : Node | None = None lowercase__ : Node | None = None class SCREAMING_SNAKE_CASE__ : def __init__( self , a): lowercase__ : Any = tree def snake_case_ ( self , a): if node is None: return 0 return node.value + ( self.depth_first_search(node.left) + self.depth_first_search(node.right) ) def __iter__( self): yield self.depth_first_search(self.tree) if __name__ == "__main__": import doctest doctest.testmod()
214
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ (__snake_case ): __lowerCamelCase : Any = """distilbert""" __lowerCamelCase : List[Any] = { """hidden_size""": """dim""", """num_attention_heads""": """n_heads""", """num_hidden_layers""": """n_layers""", } def __init__( self , a=3_0522 , a=512 , a=False , a=6 , a=12 , a=768 , a=4 * 768 , a=0.1 , a=0.1 , a="gelu" , a=0.02 , a=0.1 , a=0.2 , a=0 , **a , ): lowercase__ : int = vocab_size lowercase__ : int = max_position_embeddings lowercase__ : List[str] = sinusoidal_pos_embds lowercase__ : Tuple = n_layers lowercase__ : Optional[int] = n_heads lowercase__ : List[Any] = dim lowercase__ : Any = hidden_dim lowercase__ : Union[str, Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : List[Any] = activation lowercase__ : Union[str, Any] = initializer_range lowercase__ : Dict = qa_dropout lowercase__ : Tuple = seq_classif_dropout super().__init__(**a , pad_token_id=a) class SCREAMING_SNAKE_CASE__ (__snake_case ): @property def snake_case_ ( self): if self.task == "multiple-choice": lowercase__ : List[str] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowercase__ : Tuple = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
214
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __A = logging.get_logger(__name__) __A = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class _snake_case ( a__ ): snake_case__ = "van" def __init__( self : int , UpperCAmelCase : List[Any]=224 , UpperCAmelCase : Optional[Any]=3 , UpperCAmelCase : Dict=[7, 3, 3, 3] , UpperCAmelCase : List[str]=[4, 2, 2, 2] , UpperCAmelCase : List[str]=[64, 128, 320, 512] , UpperCAmelCase : int=[3, 3, 12, 3] , UpperCAmelCase : Optional[Any]=[8, 8, 4, 4] , UpperCAmelCase : int="gelu" , UpperCAmelCase : List[Any]=0.0_2 , UpperCAmelCase : Tuple=1E-6 , UpperCAmelCase : Dict=1E-2 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : Dict=0.0 , **UpperCAmelCase : Dict , ): super().__init__(**UpperCAmelCase ) __lowerCamelCase : Any = image_size __lowerCamelCase : List[str] = num_channels __lowerCamelCase : str = patch_sizes __lowerCamelCase : str = strides __lowerCamelCase : Optional[int] = hidden_sizes __lowerCamelCase : str = depths __lowerCamelCase : Optional[Any] = mlp_ratios __lowerCamelCase : int = hidden_act __lowerCamelCase : Optional[int] = initializer_range __lowerCamelCase : List[Any] = layer_norm_eps __lowerCamelCase : Union[str, Any] = layer_scale_init_value __lowerCamelCase : str = drop_path_rate __lowerCamelCase : Optional[Any] = dropout_rate
64
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _snake_case ( a__ ): def __init__( self : Optional[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : str ): super().__init__() # make sure scheduler can always be converted to DDIM __lowerCamelCase : Dict = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__( self : str , UpperCAmelCase : int = 1 , UpperCAmelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase : float = 0.0 , UpperCAmelCase : int = 50 , UpperCAmelCase : Optional[bool] = None , UpperCAmelCase : Optional[str] = "pil" , UpperCAmelCase : bool = True , ): # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size , UpperCAmelCase ): __lowerCamelCase : Any = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: __lowerCamelCase : Dict = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(UpperCAmelCase )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) __lowerCamelCase : str = randn_tensor(UpperCAmelCase , generator=UpperCAmelCase , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __lowerCamelCase : Any = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __lowerCamelCase : Union[str, Any] = self.scheduler.step( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , eta=UpperCAmelCase , use_clipped_model_output=UpperCAmelCase , generator=UpperCAmelCase ).prev_sample __lowerCamelCase : Any = (image / 2 + 0.5).clamp(0 , 1 ) __lowerCamelCase : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __lowerCamelCase : str = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
64
1
from __future__ import annotations def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' return len(set(UpperCamelCase__ ) ) == len(UpperCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
285
from __future__ import annotations import numpy as np def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' snake_case_ , snake_case_ = np.shape(UpperCamelCase__ ) if rows != columns: snake_case_ = ( '\'table\' has to be of square shaped array but got a ' F'''{rows}x{columns} array:\n{table}''' ) raise ValueError(UpperCamelCase__ ) snake_case_ = np.zeros((rows, columns) ) snake_case_ = np.zeros((rows, columns) ) for i in range(UpperCamelCase__ ): for j in range(UpperCamelCase__ ): snake_case_ = sum(lower[i][k] * upper[k][j] for k in range(UpperCamelCase__ ) ) if upper[j][j] == 0: raise ArithmeticError('No LU decomposition exists' ) snake_case_ = (table[i][j] - total) / upper[j][j] snake_case_ = 1 for j in range(UpperCamelCase__ , UpperCamelCase__ ): snake_case_ = sum(lower[i][k] * upper[k][j] for k in range(UpperCamelCase__ ) ) snake_case_ = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
285
1
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' UpperCAmelCase : int =mock.Mock() UpperCAmelCase : Dict =500 UpperCAmelCase : Tuple ={} UpperCAmelCase : str =HTTPError UpperCAmelCase : List[Any] ={} # Download this model to make sure it's in the cache. UpperCAmelCase : Optional[int] =BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=snake_case__ ) as mock_head: UpperCAmelCase : Dict =BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : Optional[int] =mock.Mock() UpperCAmelCase : str =500 UpperCAmelCase : Optional[Any] ={} UpperCAmelCase : Union[str, Any] =HTTPError UpperCAmelCase : List[Any] ={} # Download this model to make sure it's in the cache. UpperCAmelCase : Union[str, Any] =GPTaTokenizerFast.from_pretrained('''gpt2''' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=snake_case__ ) as mock_head: UpperCAmelCase : Union[str, Any] =GPTaTokenizerFast.from_pretrained('''gpt2''' ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' try: UpperCAmelCase : Optional[Any] =tempfile.mktemp() with open(snake_case__ , '''wb''' ) as f: http_get('''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''' , snake_case__ ) UpperCAmelCase : Optional[int] =AlbertTokenizer.from_pretrained(snake_case__ ) finally: os.remove(snake_case__ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('''tokenizer.json''' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('''tokenizer.json''' , '''wb''' ) as f: http_get('''https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json''' , snake_case__ ) UpperCAmelCase : List[Any] =AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('''tokenizer.json''' ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : Optional[Any] =AlbertTokenizer.from_pretrained('''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''' ) @is_staging_test class __snake_case ( unittest.TestCase ): __lowerCamelCase : List[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def UpperCAmelCase__ ( cls ) -> Dict: '''simple docstring''' UpperCAmelCase : str =TOKEN HfFolder.save_token(snake_case__ ) @classmethod def UpperCAmelCase__ ( cls ) -> Optional[int]: '''simple docstring''' try: delete_repo(token=cls._token , repo_id='''test-tokenizer''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-tokenizer-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-tokenizer''' ) except HTTPError: pass def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase : Any =os.path.join(snake_case__ , '''vocab.txt''' ) with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) UpperCAmelCase : Any =BertTokenizer(snake_case__ ) tokenizer.push_to_hub('''test-tokenizer''' , use_auth_token=self._token ) UpperCAmelCase : List[str] =BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='''test-tokenizer''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(snake_case__ , repo_id='''test-tokenizer''' , push_to_hub=snake_case__ , use_auth_token=self._token ) UpperCAmelCase : List[str] =BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase : Union[str, Any] =os.path.join(snake_case__ , '''vocab.txt''' ) with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) UpperCAmelCase : Optional[Any] =BertTokenizer(snake_case__ ) tokenizer.push_to_hub('''valid_org/test-tokenizer-org''' , use_auth_token=self._token ) UpperCAmelCase : Tuple =BertTokenizer.from_pretrained('''valid_org/test-tokenizer-org''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-tokenizer-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( snake_case__ , repo_id='''valid_org/test-tokenizer-org''' , push_to_hub=snake_case__ , use_auth_token=self._token ) UpperCAmelCase : str =BertTokenizer.from_pretrained('''valid_org/test-tokenizer-org''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase : Dict =os.path.join(snake_case__ , '''vocab.txt''' ) with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) UpperCAmelCase : Tuple =CustomTokenizer(snake_case__ ) # No fast custom tokenizer tokenizer.push_to_hub('''test-dynamic-tokenizer''' , use_auth_token=self._token ) UpperCAmelCase : List[str] =AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=snake_case__ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizer''' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase : List[Any] =os.path.join(snake_case__ , '''vocab.txt''' ) with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) UpperCAmelCase : Tuple =BertTokenizerFast.from_pretrained(snake_case__ ) bert_tokenizer.save_pretrained(snake_case__ ) UpperCAmelCase : Union[str, Any] =CustomTokenizerFast.from_pretrained(snake_case__ ) tokenizer.push_to_hub('''test-dynamic-tokenizer''' , use_auth_token=self._token ) UpperCAmelCase : str =AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=snake_case__ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizerFast''' ) UpperCAmelCase : Any =AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=snake_case__ , trust_remote_code=snake_case__ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizer''' ) class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Optional[Any] =Trie() trie.add('''Hello 友達''' ) self.assertEqual(trie.data , {'''H''': {'''e''': {'''l''': {'''l''': {'''o''': {''' ''': {'''友''': {'''達''': {'''''': 1}}}}}}}}} ) trie.add('''Hello''' ) trie.data self.assertEqual(trie.data , {'''H''': {'''e''': {'''l''': {'''l''': {'''o''': {'''''': 1, ''' ''': {'''友''': {'''達''': {'''''': 1}}}}}}}}} ) def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Dict =Trie() self.assertEqual(trie.split('''[CLS] This is a extra_id_100''' ) , ['''[CLS] This is a extra_id_100'''] ) trie.add('''[CLS]''' ) trie.add('''extra_id_1''' ) trie.add('''extra_id_100''' ) self.assertEqual(trie.split('''[CLS] This is a extra_id_100''' ) , ['''[CLS]''', ''' This is a ''', '''extra_id_100'''] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : Optional[int] =Trie() trie.add('''A''' ) self.assertEqual(trie.split('''ABC''' ) , ['''A''', '''BC'''] ) self.assertEqual(trie.split('''BCA''' ) , ['''BC''', '''A'''] ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : Dict =Trie() trie.add('''TOKEN]''' ) trie.add('''[SPECIAL_TOKEN]''' ) self.assertEqual(trie.split('''This is something [SPECIAL_TOKEN]''' ) , ['''This is something ''', '''[SPECIAL_TOKEN]'''] ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Optional[int] =Trie() trie.add('''A''' ) trie.add('''P''' ) trie.add('''[SPECIAL_TOKEN]''' ) self.assertEqual(trie.split('''This is something [SPECIAL_TOKEN]''' ) , ['''This is something ''', '''[SPECIAL_TOKEN]'''] ) def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' UpperCAmelCase : Dict =Trie() trie.add('''AB''' ) trie.add('''B''' ) trie.add('''C''' ) self.assertEqual(trie.split('''ABC''' ) , ['''AB''', '''C'''] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : int =Trie() trie.add('''ABC''' ) trie.add('''B''' ) trie.add('''CD''' ) self.assertEqual(trie.split('''ABCD''' ) , ['''ABC''', '''D'''] ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : str =Trie() UpperCAmelCase : Optional[int] =trie.cut_text('''ABC''' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(snake_case__ , ['''AB''', '''C'''] )
78
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class __snake_case ( unittest.TestCase ): __lowerCamelCase : Dict = MODEL_FOR_MASKED_LM_MAPPING __lowerCamelCase : Optional[int] = TF_MODEL_FOR_MASKED_LM_MAPPING def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : List[Any] =pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , top_k=2 , framework='''tf''' ) UpperCAmelCase : List[Any] =unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ {'''sequence''': '''My name is grouped''', '''score''': 2.1e-05, '''token''': 3_8015, '''token_str''': ''' grouped'''}, {'''sequence''': '''My name is accuser''', '''score''': 2.1e-05, '''token''': 2_5506, '''token_str''': ''' accuser'''}, ] , ) UpperCAmelCase : Tuple =unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ { '''sequence''': '''The largest city in France is grouped''', '''score''': 2.1e-05, '''token''': 3_8015, '''token_str''': ''' grouped''', }, { '''sequence''': '''The largest city in France is accuser''', '''score''': 2.1e-05, '''token''': 2_5506, '''token_str''': ''' accuser''', }, ] , ) UpperCAmelCase : Dict =unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ {'''sequence''': '''My name is Clara''', '''score''': 2e-05, '''token''': 1_3606, '''token_str''': ''' Clara'''}, {'''sequence''': '''My name is Patrick''', '''score''': 2e-05, '''token''': 3499, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Te''', '''score''': 1.9e-05, '''token''': 2941, '''token_str''': ''' Te'''}, ] , ) @require_torch def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' UpperCAmelCase : Any =pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , top_k=2 , framework='''pt''' ) UpperCAmelCase : List[Any] =unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ {'''sequence''': '''My name is Maul''', '''score''': 2.2e-05, '''token''': 3_5676, '''token_str''': ''' Maul'''}, {'''sequence''': '''My name isELS''', '''score''': 2.2e-05, '''token''': 1_6416, '''token_str''': '''ELS'''}, ] , ) UpperCAmelCase : Union[str, Any] =unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ { '''sequence''': '''The largest city in France is Maul''', '''score''': 2.2e-05, '''token''': 3_5676, '''token_str''': ''' Maul''', }, {'''sequence''': '''The largest city in France isELS''', '''score''': 2.2e-05, '''token''': 1_6416, '''token_str''': '''ELS'''}, ] , ) UpperCAmelCase : int =unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ {'''sequence''': '''My name is Patrick''', '''score''': 2.1e-05, '''token''': 3499, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Te''', '''score''': 2e-05, '''token''': 2941, '''token_str''': ''' Te'''}, {'''sequence''': '''My name is Clara''', '''score''': 2e-05, '''token''': 1_3606, '''token_str''': ''' Clara'''}, ] , ) UpperCAmelCase : List[str] =unmasker('''My name is <mask> <mask>''' , top_k=2 ) self.assertEqual( nested_simplify(snake_case__ , decimals=6 ) , [ [ { '''score''': 2.2e-05, '''token''': 3_5676, '''token_str''': ''' Maul''', '''sequence''': '''<s>My name is Maul<mask></s>''', }, {'''score''': 2.2e-05, '''token''': 1_6416, '''token_str''': '''ELS''', '''sequence''': '''<s>My name isELS<mask></s>'''}, ], [ { '''score''': 2.2e-05, '''token''': 3_5676, '''token_str''': ''' Maul''', '''sequence''': '''<s>My name is<mask> Maul</s>''', }, {'''score''': 2.2e-05, '''token''': 1_6416, '''token_str''': '''ELS''', '''sequence''': '''<s>My name is<mask>ELS</s>'''}, ], ] , ) @require_torch_gpu def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' UpperCAmelCase : Optional[int] =pipeline('''fill-mask''' , model='''hf-internal-testing/tiny-random-distilbert''' , device=0 , framework='''pt''' ) # convert model to fp16 pipe.model.half() UpperCAmelCase : Dict =pipe('''Paris is the [MASK] of France.''' ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(snake_case__ , snake_case__ ) @slow @require_torch def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : str =pipeline(task='''fill-mask''' , model='''distilroberta-base''' , top_k=2 , framework='''pt''' ) self.run_large_test(snake_case__ ) @slow @require_tf def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =pipeline(task='''fill-mask''' , model='''distilroberta-base''' , top_k=2 , framework='''tf''' ) self.run_large_test(snake_case__ ) def UpperCAmelCase__ ( self , snake_case__ ) -> int: '''simple docstring''' UpperCAmelCase : Union[str, Any] =unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ ) , [ {'''sequence''': '''My name is John''', '''score''': 0.008, '''token''': 610, '''token_str''': ''' John'''}, {'''sequence''': '''My name is Chris''', '''score''': 0.007, '''token''': 1573, '''token_str''': ''' Chris'''}, ] , ) UpperCAmelCase : int =unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(snake_case__ ) , [ { '''sequence''': '''The largest city in France is Paris''', '''score''': 0.251, '''token''': 2201, '''token_str''': ''' Paris''', }, { '''sequence''': '''The largest city in France is Lyon''', '''score''': 0.214, '''token''': 1_2790, '''token_str''': ''' Lyon''', }, ] , ) UpperCAmelCase : int =unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(snake_case__ ) , [ {'''sequence''': '''My name is Patrick''', '''score''': 0.005, '''token''': 3499, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Clara''', '''score''': 0.000, '''token''': 1_3606, '''token_str''': ''' Clara'''}, {'''sequence''': '''My name is Te''', '''score''': 0.000, '''token''': 2941, '''token_str''': ''' Te'''}, ] , ) @require_torch def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Any =pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , framework='''pt''' ) UpperCAmelCase : List[str] =None UpperCAmelCase : str =None self.run_pipeline_test(snake_case__ , [] ) @require_tf def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : List[str] =pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , framework='''tf''' ) UpperCAmelCase : str =None UpperCAmelCase : Dict =None self.run_pipeline_test(snake_case__ , [] ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Any: '''simple docstring''' if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest('''The provided tokenizer has no mask token, (probably reformer or wav2vec2)''' ) UpperCAmelCase : Any =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) UpperCAmelCase : int =[ f'''This is another {tokenizer.mask_token} test''', ] return fill_masker, examples def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Any =fill_masker.tokenizer UpperCAmelCase : Optional[int] =fill_masker.model UpperCAmelCase : Dict =fill_masker( f'''This is a {tokenizer.mask_token}''' , ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) UpperCAmelCase : int =fill_masker([f'''This is a {tokenizer.mask_token}'''] ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) UpperCAmelCase : Optional[Any] =fill_masker([f'''This is a {tokenizer.mask_token}''', f'''Another {tokenizer.mask_token} great test.'''] ) self.assertEqual( snake_case__ , [ [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ], [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ], ] , ) with self.assertRaises(snake_case__ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(snake_case__ ): fill_masker('''This is''' ) self.run_test_top_k(snake_case__ , snake_case__ ) self.run_test_targets(snake_case__ , snake_case__ ) self.run_test_top_k_targets(snake_case__ , snake_case__ ) self.fill_mask_with_duplicate_targets_and_top_k(snake_case__ , snake_case__ ) self.fill_mask_with_multiple_masks(snake_case__ , snake_case__ ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> List[str]: '''simple docstring''' UpperCAmelCase : List[str] =tokenizer.get_vocab() UpperCAmelCase : List[str] =sorted(vocab.keys() )[:2] # Pipeline argument UpperCAmelCase : Any =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ , targets=snake_case__ ) UpperCAmelCase : Tuple =fill_masker(f'''This is a {tokenizer.mask_token}''' ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) UpperCAmelCase : int ={vocab[el] for el in targets} self.assertEqual({el['''token'''] for el in outputs} , snake_case__ ) UpperCAmelCase : Optional[Any] =[tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['''token_str'''] for el in outputs} , set(snake_case__ ) ) # Call argument UpperCAmelCase : Union[str, Any] =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) UpperCAmelCase : Optional[int] =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets=snake_case__ ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) UpperCAmelCase : Any ={vocab[el] for el in targets} self.assertEqual({el['''token'''] for el in outputs} , snake_case__ ) UpperCAmelCase : Optional[int] =[tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['''token_str'''] for el in outputs} , set(snake_case__ ) ) # Score equivalence UpperCAmelCase : Any =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets=snake_case__ ) UpperCAmelCase : Union[str, Any] =[top_mask['''token_str'''] for top_mask in outputs] UpperCAmelCase : Optional[int] =[top_mask['''score'''] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(snake_case__ ) == set(snake_case__ ): UpperCAmelCase : List[Any] =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets=snake_case__ ) UpperCAmelCase : List[str] =[top_mask['''score'''] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(snake_case__ ) , nested_simplify(snake_case__ ) ) # Raises with invalid with self.assertRaises(snake_case__ ): UpperCAmelCase : int =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(snake_case__ ): UpperCAmelCase : int =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets=[''''''] ) with self.assertRaises(snake_case__ ): UpperCAmelCase : Any =fill_masker(f'''This is a {tokenizer.mask_token}''' , targets='''''' ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> Tuple: '''simple docstring''' UpperCAmelCase : List[str] =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ , top_k=2 ) UpperCAmelCase : Optional[Any] =fill_masker(f'''This is a {tokenizer.mask_token}''' ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) UpperCAmelCase : Any =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) UpperCAmelCase : List[Any] =fill_masker(f'''This is a {tokenizer.mask_token}''' , top_k=2 ) self.assertEqual( snake_case__ , [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ] , ) self.assertEqual(nested_simplify(snake_case__ ) , nested_simplify(snake_case__ ) ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : str =tokenizer.get_vocab() UpperCAmelCase : int =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) # top_k=2, ntargets=3 UpperCAmelCase : Optional[Any] =sorted(vocab.keys() )[:3] UpperCAmelCase : str =fill_masker(f'''This is a {tokenizer.mask_token}''' , top_k=2 , targets=snake_case__ ) # If we use the most probably targets, and filter differently, we should still # have the same results UpperCAmelCase : Tuple =[el['''token_str'''] for el in sorted(snake_case__ , key=lambda snake_case__ : x["score"] , reverse=snake_case__ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(snake_case__ ).issubset(snake_case__ ): UpperCAmelCase : str =fill_masker(f'''This is a {tokenizer.mask_token}''' , top_k=3 , targets=snake_case__ ) # They should yield exactly the same result self.assertEqual(nested_simplify(snake_case__ ) , nested_simplify(snake_case__ ) ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> Dict: '''simple docstring''' UpperCAmelCase : str =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) UpperCAmelCase : Union[str, Any] =tokenizer.get_vocab() # String duplicates + id duplicates UpperCAmelCase : List[Any] =sorted(vocab.keys() )[:3] UpperCAmelCase : Optional[int] =[targets[0], targets[1], targets[0], targets[2], targets[1]] UpperCAmelCase : str =fill_masker(f'''My name is {tokenizer.mask_token}''' , targets=snake_case__ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(snake_case__ ) , 3 ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Any =FillMaskPipeline(model=snake_case__ , tokenizer=snake_case__ ) UpperCAmelCase : Union[str, Any] =fill_masker( f'''This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}''' , top_k=2 ) self.assertEqual( snake_case__ , [ [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ], [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ], [ {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, {'''sequence''': ANY(snake_case__ ), '''score''': ANY(snake_case__ ), '''token''': ANY(snake_case__ ), '''token_str''': ANY(snake_case__ )}, ], ] , )
78
1
'''simple docstring''' def _SCREAMING_SNAKE_CASE ( UpperCamelCase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" lowerCAmelCase__ : str = set() # Replace all the whitespace in our sentence lowerCAmelCase__ : Tuple = input_str.replace(""" """ , """""" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(UpperCamelCase ) == 26 def _SCREAMING_SNAKE_CASE ( UpperCamelCase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" lowerCAmelCase__ : Any = [False] * 26 for char in input_str: if char.islower(): lowerCAmelCase__ : Optional[Any] = True elif char.isupper(): lowerCAmelCase__ : Any = True return all(UpperCamelCase ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" from timeit import timeit lowerCAmelCase__ : Union[str, Any] = """from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest""" print(timeit("""is_pangram()""" , setup=UpperCamelCase ) ) print(timeit("""is_pangram_faster()""" , setup=UpperCamelCase ) ) print(timeit("""is_pangram_fastest()""" , setup=UpperCamelCase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
37
'''simple docstring''' import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = {'''vocab_file''': '''spiece.model'''} _lowerCAmelCase = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', } } _lowerCAmelCase = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } _lowerCAmelCase = '''▁''' class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : Any = VOCAB_FILES_NAMES __lowercase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __lowercase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=False ,__UpperCAmelCase="[CLS]" ,__UpperCAmelCase="[SEP]" ,__UpperCAmelCase="<unk>" ,__UpperCAmelCase="[SEP]" ,__UpperCAmelCase="<pad>" ,__UpperCAmelCase="[CLS]" ,__UpperCAmelCase="[MASK]" ,__UpperCAmelCase = None ,**__UpperCAmelCase ,) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCAmelCase__ : Tuple = ( AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ,normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else mask_token ) lowerCAmelCase__ : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase ,remove_space=__UpperCAmelCase ,keep_accents=__UpperCAmelCase ,bos_token=__UpperCAmelCase ,eos_token=__UpperCAmelCase ,unk_token=__UpperCAmelCase ,sep_token=__UpperCAmelCase ,pad_token=__UpperCAmelCase ,cls_token=__UpperCAmelCase ,mask_token=__UpperCAmelCase ,sp_model_kwargs=self.sp_model_kwargs ,**__UpperCAmelCase ,) lowerCAmelCase__ : str = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : Tuple = keep_accents lowerCAmelCase__ : Any = vocab_file lowerCAmelCase__ : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def UpperCAmelCase_ ( self ) -> Optional[int]: return len(self.sp_model ) def UpperCAmelCase_ ( self ) -> Optional[Any]: lowerCAmelCase__ : Union[str, Any] = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Any: lowerCAmelCase__ : Optional[Any] = self.__dict__.copy() lowerCAmelCase__ : Optional[Any] = None return state def __setstate__( self ,__UpperCAmelCase ) -> List[Any]: lowerCAmelCase__ : List[str] = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): lowerCAmelCase__ : Union[str, Any] = {} lowerCAmelCase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Optional[int]: if self.remove_space: lowerCAmelCase__ : int = """ """.join(inputs.strip().split() ) else: lowerCAmelCase__ : str = inputs lowerCAmelCase__ : Tuple = outputs.replace("""``""" ,"""\"""" ).replace("""''""" ,"""\"""" ) if not self.keep_accents: lowerCAmelCase__ : Any = unicodedata.normalize("""NFKD""" ,__UpperCAmelCase ) lowerCAmelCase__ : Dict = """""".join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: lowerCAmelCase__ : Tuple = outputs.lower() return outputs def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> List[str]: lowerCAmelCase__ : List[str] = self.preprocess_text(__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = self.sp_model.encode(__UpperCAmelCase ,out_type=__UpperCAmelCase ) lowerCAmelCase__ : str = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowerCAmelCase__ : List[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase ,"""""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCAmelCase__ : str = cur_pieces[1:] else: lowerCAmelCase__ : int = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> List[str]: return self.sp_model.PieceToId(__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Any: return self.sp_model.IdToPiece(__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> str: lowerCAmelCase__ : str = [] lowerCAmelCase__ : Tuple = """""" lowerCAmelCase__ : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : List[str] = [] else: current_sub_tokens.append(__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> List[int]: lowerCAmelCase__ : int = [self.sep_token_id] lowerCAmelCase__ : Dict = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ,__UpperCAmelCase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase ,token_ids_a=__UpperCAmelCase ,already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> List[int]: lowerCAmelCase__ : List[str] = [self.sep_token_id] lowerCAmelCase__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> Tuple[str]: if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase__ : int = os.path.join( __UpperCAmelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,__UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase ,"""wb""" ) as fi: lowerCAmelCase__ : List[Any] = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
37
1
"""simple docstring""" from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def lowerCamelCase_( _lowerCamelCase ) -> Optional[int]: '''simple docstring''' _lowerCamelCase : Dict = [] _lowerCamelCase : List[Any] = [] _lowerCamelCase : List[Any] = [] for rt in rc.restypes: _lowerCamelCase : Optional[int] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) _lowerCamelCase : Any = {name: i for i, name in enumerate(a__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) _lowerCamelCase : Optional[Any] = torch.tensor( a__ , dtype=torch.intaa , device=protein["aatype"].device , ) _lowerCamelCase : Tuple = torch.tensor( a__ , dtype=torch.intaa , device=protein["aatype"].device , ) _lowerCamelCase : Dict = torch.tensor( a__ , dtype=torch.floataa , device=protein["aatype"].device , ) _lowerCamelCase : Optional[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein _lowerCamelCase : Optional[Any] = restype_atomaa_to_atomaa[protein_aatype] _lowerCamelCase : Dict = restype_atomaa_mask[protein_aatype] _lowerCamelCase : List[Any] = residx_atomaa_mask _lowerCamelCase : Dict = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back _lowerCamelCase : Dict = restype_atomaa_to_atomaa[protein_aatype] _lowerCamelCase : str = residx_atomaa_to_atomaa.long() # create the corresponding mask _lowerCamelCase : Optional[int] = torch.zeros([21, 37] , dtype=torch.floataa , device=protein["aatype"].device ) for restype, restype_letter in enumerate(rc.restypes ): _lowerCamelCase : Tuple = rc.restype_atoa[restype_letter] _lowerCamelCase : Union[str, Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: _lowerCamelCase : Tuple = rc.atom_order[atom_name] _lowerCamelCase : Any = 1 _lowerCamelCase : Tuple = restype_atomaa_mask[protein_aatype] _lowerCamelCase : Dict = residx_atomaa_mask return protein def lowerCamelCase_( _lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase : Union[str, Any] = tree_map(lambda _lowerCamelCase : torch.tensor(a__ , device=batch["aatype"].device ) , a__ , np.ndarray ) _lowerCamelCase : int = tensor_tree_map(lambda _lowerCamelCase : np.array(a__ ) , make_atomaa_masks(a__ ) ) return out
369
"""simple docstring""" def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> str: '''simple docstring''' _lowerCamelCase : int = len(_lowerCamelCase ) _lowerCamelCase : int = len(_lowerCamelCase ) _lowerCamelCase : int = ( first_str_length if first_str_length > second_str_length else second_str_length ) _lowerCamelCase : list = [] for char_count in range(_lowerCamelCase ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(_lowerCamelCase ) if __name__ == "__main__": print(alternative_string_arrange('''AB''', '''XYZ'''), end=''' ''')
340
0
import datasets from .evaluate import evaluate a_ = "\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n" a_ = "\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n" a_ = "\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the SQuAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}]\n >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}]\n >>> squad_metric = datasets.load_metric(\"squad\")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': {'id': datasets.Value('string' ), 'prediction_text': datasets.Value('string' )}, 'references': { 'id': datasets.Value('string' ), 'answers': datasets.features.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), }, } ) , codebase_urls=['https://rajpurkar.github.io/SQuAD-explorer/'] , reference_urls=['https://rajpurkar.github.io/SQuAD-explorer/'] , ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , snake_case : Any , snake_case : Any ) -> str: """simple docstring""" UpperCamelCase_ : str = {prediction['id']: prediction['prediction_text'] for prediction in predictions} UpperCamelCase_ : int = [ { 'paragraphs': [ { 'qas': [ { 'answers': [{'text': answer_text} for answer_text in ref['answers']['text']], 'id': ref['id'], } for ref in references ] } ] } ] UpperCamelCase_ : List[str] = evaluate(dataset=lowerCamelCase_ , predictions=lowerCamelCase_ ) return score
175
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def a( A : dict ) -> tuple: """simple docstring""" return (data["data"], data["target"]) def a( A : np.ndarray , A : np.ndarray ) -> XGBClassifier: """simple docstring""" a = XGBClassifier() classifier.fit(A , A ) return classifier def a( ) -> None: """simple docstring""" a = load_iris() a , a = data_handling(A ) a , a , a , a = train_test_split( A , A , test_size=0.25 ) a = iris["target_names"] # Create an XGBoost Classifier from the training data a = xgboost(A , A ) # Display the confusion matrix of the classifier with both training and test sets ConfusionMatrixDisplay.from_estimator( A , A , A , display_labels=A , cmap="Blues" , normalize="true" , ) plt.title("Normalized Confusion Matrix - IRIS Dataset" ) plt.show() if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
227
0
'''simple docstring''' def _lowerCAmelCase ( __snake_case : str , __snake_case : str ) -> str: __A : int = len(__snake_case ) __A : int = len(__snake_case ) __A : int = ( first_str_length if first_str_length > second_str_length else second_str_length ) __A : list = [] for char_count in range(__snake_case ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(__snake_case ) if __name__ == "__main__": print(alternative_string_arrange('''AB''', '''XYZ'''), end=''' ''')
190
'''simple docstring''' import itertools import math def _lowerCAmelCase ( __snake_case : int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _lowerCAmelCase ( ) -> List[Any]: __A : Optional[Any] = 2 while True: if is_prime(__snake_case ): yield num num += 1 def _lowerCAmelCase ( __snake_case : int = 1_00_01 ) -> int: return next(itertools.islice(prime_generator() , nth - 1 , __snake_case ) ) if __name__ == "__main__": print(f"""{solution() = }""")
190
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __a = logging.get_logger(__name__) __a = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __a = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } __a = { "distilbert-base-uncased": 5_12, "distilbert-base-uncased-distilled-squad": 5_12, "distilbert-base-cased": 5_12, "distilbert-base-cased-distilled-squad": 5_12, "distilbert-base-german-cased": 5_12, "distilbert-base-multilingual-cased": 5_12, } __a = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( _lowerCAmelCase ): '''simple docstring''' _A : int = VOCAB_FILES_NAMES _A : List[Any] = PRETRAINED_VOCAB_FILES_MAP _A : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : List[Any] = PRETRAINED_INIT_CONFIGURATION _A : int = ["""input_ids""", """attention_mask"""] _A : str = DistilBertTokenizer def __init__( self: List[Any] , snake_case: Dict=None , snake_case: Tuple=None , snake_case: List[str]=True , snake_case: int="[UNK]" , snake_case: List[str]="[SEP]" , snake_case: Any="[PAD]" , snake_case: List[Any]="[CLS]" , snake_case: int="[MASK]" , snake_case: Optional[Any]=True , snake_case: Dict=None , **snake_case: List[str] , ) -> Optional[int]: super().__init__( snake_case , tokenizer_file=snake_case , do_lower_case=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , tokenize_chinese_chars=snake_case , strip_accents=snake_case , **snake_case , ) snake_case_ :str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , snake_case ) != do_lower_case or normalizer_state.get("""strip_accents""" , snake_case ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , snake_case ) != tokenize_chinese_chars ): snake_case_ :Optional[Any] = getattr(snake_case , normalizer_state.pop("""type""" ) ) snake_case_ :Any = do_lower_case snake_case_ :Optional[int] = strip_accents snake_case_ :Optional[Any] = tokenize_chinese_chars snake_case_ :Dict = normalizer_class(**snake_case ) snake_case_ :Optional[Any] = do_lower_case def lowerCAmelCase_ ( self: Dict , snake_case: Union[str, Any] , snake_case: Union[str, Any]=None ) -> int: snake_case_ :Union[str, Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase_ ( self: Any , snake_case: List[int] , snake_case: Optional[List[int]] = None ) -> List[int]: snake_case_ :List[Any] = [self.sep_token_id] snake_case_ :List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase_ ( self: str , snake_case: str , snake_case: Optional[str] = None ) -> Tuple[str]: snake_case_ :Any = self._tokenizer.model.save(snake_case , name=snake_case ) return tuple(snake_case )
66
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): # initialize config if "resnet-50" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-50") elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-101") else: raise ValueError("Model name should include either resnet50 or resnet101") SCREAMING_SNAKE_CASE : str = DetrConfig(use_timm_backbone=_a , backbone_config=_a) # set label attributes SCREAMING_SNAKE_CASE : List[str] = "panoptic" in model_name if is_panoptic: SCREAMING_SNAKE_CASE : Union[str, Any] = 250 else: SCREAMING_SNAKE_CASE : Union[str, Any] = 91 SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : Union[str, Any] = "coco-detection-id2label.json" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(_a , _a , repo_type="dataset") , "r")) SCREAMING_SNAKE_CASE : int = {int(_a): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCamelCase__ ( _a): # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE : Union[str, Any] = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight")) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight")) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias")) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean")) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var")) # stages for stage_idx in range(len(config.backbone_config.depths)): for layer_idx in range(config.backbone_config.depths[stage_idx]): # shortcut if layer_idx == 0: rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var", )) # 3 convs for i in range(3): rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var", )) # fmt: on for i in range(config.encoder_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", )) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", )) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ]) return rename_keys def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : str = state_dict.pop(_a) SCREAMING_SNAKE_CASE : int = val def lowerCamelCase__ ( _a , _a=False): SCREAMING_SNAKE_CASE : Optional[Any] = "" if is_panoptic: SCREAMING_SNAKE_CASE : Optional[int] = "detr." # first: transformer encoder for i in range(6): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : int = in_proj_bias[:256] SCREAMING_SNAKE_CASE : Tuple = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : str = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : str = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[:256] SCREAMING_SNAKE_CASE : List[Any] = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : Any = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight") SCREAMING_SNAKE_CASE : int = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias") # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE : Tuple = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[-256:] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Union[str, Any] = Image.open(requests.get(_a , stream=_a).raw) return im @torch.no_grad() def lowerCamelCase__ ( _a , _a=None , _a=False): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Optional[int] = get_detr_config(_a) # load original model from torch hub SCREAMING_SNAKE_CASE : Union[str, Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f"Converting model {model_name}...") SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=_a).eval() SCREAMING_SNAKE_CASE : Tuple = detr.state_dict() # rename keys for src, dest in create_rename_keys(_a): if is_panoptic: SCREAMING_SNAKE_CASE : List[str] = "detr." + src rename_key(_a , _a , _a) # query, key and value matrices need special treatment read_in_q_k_v(_a , is_panoptic=_a) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE : List[Any] = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr") and not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor") ): SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Optional[int] = val elif key.startswith("bbox_attention") or key.startswith("mask_head"): continue else: SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : List[Any] = val else: if not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor"): SCREAMING_SNAKE_CASE : Any = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Any = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE : int = DetrForSegmentation(_a) if is_panoptic else DetrForObjectDetection(_a) model.load_state_dict(_a) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE : int = "coco_panoptic" if is_panoptic else "coco_detection" SCREAMING_SNAKE_CASE : Optional[int] = DetrImageProcessor(format=_a) SCREAMING_SNAKE_CASE : List[str] = processor(images=prepare_img() , return_tensors="pt") SCREAMING_SNAKE_CASE : Any = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = detr(_a) SCREAMING_SNAKE_CASE : Any = model(_a) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4) print("Looks ok!") if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}...") Path(_a).mkdir(exist_ok=_a) model.save_pretrained(_a) processor.save_pretrained(_a) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub...") model.push_to_hub(f"nielsr/{model_name}") processor.push_to_hub(f"nielsr/{model_name}") if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') a_ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
76
0
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def UpperCamelCase ( _a ) -> Any: '''simple docstring''' return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def UpperCamelCase ( _a ) -> Optional[int]: '''simple docstring''' lowercase_ :int = create_tensor(_a ) lowercase_ :Optional[int] = gather(_a ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def UpperCamelCase ( _a ) -> str: '''simple docstring''' lowercase_ :Tuple = [state.process_index] lowercase_ :Dict = gather_object(_a ) assert len(_a ) == state.num_processes, f"{gathered_obj}, {len(_a )} != {state.num_processes}" assert gathered_obj == list(range(state.num_processes ) ), f"{gathered_obj} != {list(range(state.num_processes ) )}" def UpperCamelCase ( _a ) -> List[str]: '''simple docstring''' lowercase_ :Union[str, Any] = create_tensor(_a ) lowercase_ :Union[str, Any] = broadcast(_a ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def UpperCamelCase ( _a ) -> Union[str, Any]: '''simple docstring''' if state.is_main_process: lowercase_ :Any = torch.arange(state.num_processes + 1 ).to(state.device ) else: lowercase_ :Optional[Any] = torch.arange(state.num_processes ).to(state.device ) lowercase_ :Optional[Any] = pad_across_processes(_a ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def UpperCamelCase ( _a ) -> Dict: '''simple docstring''' if state.num_processes != 2: return lowercase_ :int = create_tensor(_a ) lowercase_ :Dict = reduce(_a , '''sum''' ) lowercase_ :str = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(_a , _a ), f"{reduced_tensor} != {truth_tensor}" def UpperCamelCase ( _a ) -> List[Any]: '''simple docstring''' if state.num_processes != 2: return lowercase_ :Optional[int] = create_tensor(_a ) lowercase_ :Optional[Any] = reduce(_a , '''mean''' ) lowercase_ :Dict = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(_a , _a ), f"{reduced_tensor} != {truth_tensor}" def UpperCamelCase ( _a ) -> str: '''simple docstring''' main() def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' lowercase_ :str = PartialState() state.print(f"State: {state}" ) state.print('''testing gather''' ) test_gather(_a ) state.print('''testing gather_object''' ) test_gather_object(_a ) state.print('''testing broadcast''' ) test_broadcast(_a ) state.print('''testing pad_across_processes''' ) test_pad_across_processes(_a ) state.print('''testing reduce_sum''' ) test_reduce_sum(_a ) state.print('''testing reduce_mean''' ) test_reduce_mean(_a ) if __name__ == "__main__": main()
366
import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase ( self ): lowercase_ :int = logging.get_logger() # the current default level is logging.WARNING lowercase_ :List[str] = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(UpperCamelCase_ ) def UpperCamelCase ( self ): lowercase_ :Tuple = logging.get_verbosity() lowercase_ :str = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Tuple = '''Testing 1, 2, 3''' # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , '''''' ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) # restore to the original level logging.set_verbosity(UpperCamelCase_ ) @mockenv(TRANSFORMERS_VERBOSITY='''error''' ) def UpperCamelCase ( self ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() # this action activates the env var lowercase_ :Any = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Optional[Any] = os.getenv('''TRANSFORMERS_VERBOSITY''' , UpperCamelCase_ ) lowercase_ :Any = logging.log_levels[env_level_str] lowercase_ :Optional[int] = logging.get_verbosity() self.assertEqual( UpperCamelCase_ , UpperCamelCase_ , f"TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}" , ) # restore to the original level lowercase_ :str = '''''' transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY='''super-error''' ) def UpperCamelCase ( self ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() lowercase_ :Any = logging.logging.getLogger() with CaptureLogger(UpperCamelCase_ ) as cl: # this action activates the env var logging.get_logger('''transformers.models.bart.tokenization_bart''' ) self.assertIn('''Unknown option TRANSFORMERS_VERBOSITY=super-error''' , cl.out ) # no need to restore as nothing was changed def UpperCamelCase ( self ): # testing `logger.warning_advice()` transformers.utils.logging._reset_library_root_logger() lowercase_ :Optional[int] = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) lowercase_ :Any = '''Testing 1, 2, 3''' with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''1''' ): # nothing should be logged as env var disables this method with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning_advice(UpperCamelCase_ ) self.assertEqual(cl.out , '''''' ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''''' ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(UpperCamelCase_ ) as cl: logger.warning_advice(UpperCamelCase_ ) self.assertEqual(cl.out , msg + '''\n''' ) def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
252
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { '''configuration_rembert''': ['''REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RemBertConfig''', '''RemBertOnnxConfig'''] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['''RemBertTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['''RemBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ '''REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RemBertForCausalLM''', '''RemBertForMaskedLM''', '''RemBertForMultipleChoice''', '''RemBertForQuestionAnswering''', '''RemBertForSequenceClassification''', '''RemBertForTokenClassification''', '''RemBertLayer''', '''RemBertModel''', '''RemBertPreTrainedModel''', '''load_tf_weights_in_rembert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ '''TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRemBertForCausalLM''', '''TFRemBertForMaskedLM''', '''TFRemBertForMultipleChoice''', '''TFRemBertForQuestionAnswering''', '''TFRemBertForSequenceClassification''', '''TFRemBertForTokenClassification''', '''TFRemBertLayer''', '''TFRemBertModel''', '''TFRemBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
181
'''simple docstring''' import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class lowerCamelCase_ ( unittest.TestCase ): def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() UpperCAmelCase__ : str = dict(zip(_A , range(len(_A ) ) ) ) UpperCAmelCase__ : int = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } UpperCAmelCase__ : List[Any] = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 16_000, '''return_attention_mask''': False, '''do_normalize''': True, } UpperCAmelCase__ : str = tempfile.mkdtemp() UpperCAmelCase__ : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) UpperCAmelCase__ : List[Any] = os.path.join(self.tmpdirname , _A ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_A ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_A ) + '''\n''' ) # load decoder from hub UpperCAmelCase__ : Tuple = '''hf-internal-testing/ngram-beam-search-decoder''' def lowercase_ ( self : Union[str, Any] , **_A : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.add_kwargs_tokens_map.copy() kwargs.update(_A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **_A ) def lowercase_ ( self : Dict , **_A : str ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **_A ) def lowercase_ ( self : Tuple , **_A : int ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **_A ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Dict = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , _A ) def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def lowercase_ ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(_A , '''include''' ): WavaVecaProcessorWithLM( tokenizer=_A , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.get_feature_extractor() UpperCAmelCase__ : List[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : str = feature_extractor(_A , return_tensors='''np''' ) UpperCAmelCase__ : Optional[Any] = processor(_A , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCAmelCase__ : Any = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Union[str, Any] = '''This is a test string''' UpperCAmelCase__ : Union[str, Any] = processor(text=_A ) UpperCAmelCase__ : Dict = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowercase_ ( self : str , _A : int=(2, 10, 16) , _A : Optional[int]=77 ): '''simple docstring''' np.random.seed(_A ) return np.random.rand(*_A ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Dict = self.get_tokenizer() UpperCAmelCase__ : Tuple = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits(shape=(10, 16) , seed=13 ) UpperCAmelCase__ : Union[str, Any] = processor.decode(_A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(_A )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def lowercase_ ( self : Dict , _A : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(_A ) else: with get_context(_A ).Pool() as pool: UpperCAmelCase__ : List[Any] = processor.batch_decode(_A , _A ) UpperCAmelCase__ : List[str] = list(_A ) with get_context('''fork''' ).Pool() as p: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch(_A , _A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(_A , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(_A , decoded_processor.logit_score ) self.assertListEqual(_A , decoded_processor.lm_score ) def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : List[Any] = self._get_dummy_logits() UpperCAmelCase__ : Optional[Any] = 15 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : List[str] = -4.0 UpperCAmelCase__ : str = processor.batch_decode( _A , beam_width=_A , beam_prune_logp=_A , token_min_logp=_A , ) UpperCAmelCase__ : Union[str, Any] = decoded_processor_out.text UpperCAmelCase__ : Tuple = list(_A ) with get_context('''fork''' ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( _A , _A , beam_width=_A , beam_prune_logp=_A , token_min_logp=_A , ) UpperCAmelCase__ : Optional[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : Tuple = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(_A , _A ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , _A ) self.assertTrue(np.array_equal(_A , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] , _A , atol=1e-3 ) ) self.assertTrue(np.array_equal(_A , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] , _A , atol=1e-3 ) ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_tokenizer() UpperCAmelCase__ : List[str] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() UpperCAmelCase__ : List[Any] = 2.0 UpperCAmelCase__ : Union[str, Any] = 5.0 UpperCAmelCase__ : Any = -2_0.0 UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : Any = processor.batch_decode( _A , alpha=_A , beta=_A , unk_score_offset=_A , lm_score_boundary=_A , ) UpperCAmelCase__ : List[Any] = decoded_processor_out.text UpperCAmelCase__ : List[Any] = list(_A ) decoder.reset_params( alpha=_A , beta=_A , unk_score_offset=_A , lm_score_boundary=_A , ) with get_context('''fork''' ).Pool() as pool: UpperCAmelCase__ : Any = decoder.decode_beams_batch( _A , _A , ) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(_A , _A ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , _A ) UpperCAmelCase__ : Any = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -2_0.0 ) self.assertEqual(lm_model.score_boundary , _A ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : List[Any] = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[Any] = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() UpperCAmelCase__ : Dict = os.listdir(_A ) UpperCAmelCase__ : str = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(_A , _A ) def lowercase_ ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = snapshot_download('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained(_A ) UpperCAmelCase__ : Optional[Any] = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Union[str, Any] = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[Any] = os.listdir(_A ) UpperCAmelCase__ : int = os.listdir(_A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(_A , _A ) def lowercase_ ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Any = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = floats_list((3, 1_000) ) UpperCAmelCase__ : Tuple = processor_wavaveca(_A , return_tensors='''np''' ) UpperCAmelCase__ : Union[str, Any] = processor_auto(_A , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) UpperCAmelCase__ : str = self._get_dummy_logits() UpperCAmelCase__ : List[Any] = processor_wavaveca.batch_decode(_A ) UpperCAmelCase__ : List[Any] = processor_auto.batch_decode(_A ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def lowercase_ ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Any = self.get_tokenizer() UpperCAmelCase__ : List[str] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def lowercase_ ( _A : Tuple , _A : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = self._get_dummy_logits()[0] UpperCAmelCase__ : int = processor.decode(_A , output_word_offsets=_A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(_A , _A ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Dict = processor.batch_decode(_A , output_word_offsets=_A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(_A , _A ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(_A , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def lowercase_ ( self : str ): '''simple docstring''' import torch UpperCAmelCase__ : Dict = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=_A ) UpperCAmelCase__ : Optional[Any] = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Any = iter(_A ) UpperCAmelCase__ : Dict = next(_A ) UpperCAmelCase__ : Optional[int] = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) UpperCAmelCase__ : List[Any] = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Union[str, Any] = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): UpperCAmelCase__ : Dict = model(_A ).logits.cpu().numpy() UpperCAmelCase__ : List[str] = processor.decode(logits[0] , output_word_offsets=_A ) UpperCAmelCase__ : Tuple = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : List[str] = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] UpperCAmelCase__ : Tuple = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(_A , '''word''' ) ) , _A ) self.assertEqual(''' '''.join(self.get_from_offsets(_A , '''word''' ) ) , output.text ) # output times UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(_A , '''start_time''' ) ) UpperCAmelCase__ : Dict = torch.tensor(self.get_from_offsets(_A , '''end_time''' ) ) # fmt: off UpperCAmelCase__ : Any = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : Any = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(_A , _A , atol=0.0_1 ) ) self.assertTrue(torch.allclose(_A , _A , atol=0.0_1 ) )
181
1
from timeit import timeit snake_case : Tuple = { '''MALAYALAM''': True, '''String''': False, '''rotor''': True, '''level''': True, '''A''': True, '''BB''': True, '''ABC''': False, '''amanaplanacanalpanama''': True, # "a man a plan a canal panama" } # Ensure our test data is valid assert all((key == key[::-1]) is value for key, value in test_data.items()) def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" a :Union[str, Any] = 0 a :Tuple = len(UpperCAmelCase_ ) - 1 while start_i < end_i: if s[start_i] == s[end_i]: start_i += 1 end_i -= 1 else: return False return True def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" a :str = len(UpperCAmelCase_ ) // 2 a :List[str] = len(UpperCAmelCase_ ) # We need to traverse till half of the length of string # as we can get access of the i'th last element from # i'th index. # eg: [0,1,2,3,4,5] => 4th index can be accessed # with the help of 1st index (i==n-i-1) # where n is length of string return all(s[i] == s[n - i - 1] for i in range(UpperCAmelCase_ ) ) def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" if len(UpperCAmelCase_ ) <= 2: return True if s[0] == s[len(UpperCAmelCase_ ) - 1]: return is_palindrome_recursive(s[1:-1] ) else: return False def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" return s == s[::-1] def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" a :Optional[int] = F'''all({name}(key) is value for key, value in test_data.items())''' a :str = F'''from __main__ import test_data, {name}''' a :List[Any] = 50_0000 a :Tuple = timeit(stmt=UpperCAmelCase_ , setup=UpperCAmelCase_ , number=UpperCAmelCase_ ) print(F'''{name:<35} finished {number:,} runs in {result:.5f} seconds''' ) if __name__ == "__main__": for key, value in test_data.items(): assert is_palindrome(key) is is_palindrome_recursive(key) assert is_palindrome(key) is is_palindrome_slice(key) print(F"""{key:21} {value}""") print('''a man a plan a canal panama''') # finished 500,000 runs in 0.46793 seconds benchmark_function('''is_palindrome_slice''') # finished 500,000 runs in 0.85234 seconds benchmark_function('''is_palindrome''') # finished 500,000 runs in 1.32028 seconds benchmark_function('''is_palindrome_recursive''') # finished 500,000 runs in 2.08679 seconds benchmark_function('''is_palindrome_traversal''')
353
from ...configuration_utils import PretrainedConfig class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = 'bert-generation' def __init__( self , _lowerCamelCase=5_0358 , _lowerCamelCase=1024 , _lowerCamelCase=24 , _lowerCamelCase=16 , _lowerCamelCase=4096 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=0.02 , _lowerCamelCase=1e-12 , _lowerCamelCase=0 , _lowerCamelCase=2 , _lowerCamelCase=1 , _lowerCamelCase="absolute" , _lowerCamelCase=True , **_lowerCamelCase , ): super().__init__(pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , **_lowerCamelCase ) a :Optional[int] = vocab_size a :Tuple = hidden_size a :Any = num_hidden_layers a :Any = num_attention_heads a :List[Any] = hidden_act a :Tuple = intermediate_size a :Any = hidden_dropout_prob a :int = attention_probs_dropout_prob a :Dict = max_position_embeddings a :int = initializer_range a :Union[str, Any] = layer_norm_eps a :str = position_embedding_type a :int = use_cache
281
0
"""simple docstring""" import os from math import logaa def _A ( UpperCamelCase_ : str = "base_exp.txt") -> int: '''simple docstring''' __lowercase = 0 __lowercase = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(UpperCamelCase_), UpperCamelCase_))): __lowercase ,__lowercase = list(map(UpperCamelCase_, line.split(","))) if x * logaa(UpperCamelCase_) > largest: __lowercase = x * logaa(UpperCamelCase_) __lowercase = i + 1 return result if __name__ == "__main__": print(solution())
17
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'huggingface/time-series-transformer-tourism-monthly': ( 'https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class _lowerCAmelCase ( lowercase ): """simple docstring""" __UpperCAmelCase : int = "time_series_transformer" __UpperCAmelCase : Any = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self : int, UpperCAmelCase__ : Optional[int] = None, UpperCAmelCase__ : Optional[int] = None, UpperCAmelCase__ : str = "student_t", UpperCAmelCase__ : str = "nll", UpperCAmelCase__ : int = 1, UpperCAmelCase__ : List[int] = [1, 2, 3, 4, 5, 6, 7], UpperCAmelCase__ : Optional[Union[str, bool]] = "mean", UpperCAmelCase__ : int = 0, UpperCAmelCase__ : int = 0, UpperCAmelCase__ : int = 0, UpperCAmelCase__ : int = 0, UpperCAmelCase__ : Optional[List[int]] = None, UpperCAmelCase__ : Optional[List[int]] = None, UpperCAmelCase__ : int = 3_2, UpperCAmelCase__ : int = 3_2, UpperCAmelCase__ : int = 2, UpperCAmelCase__ : int = 2, UpperCAmelCase__ : int = 2, UpperCAmelCase__ : int = 2, UpperCAmelCase__ : bool = True, UpperCAmelCase__ : str = "gelu", UpperCAmelCase__ : int = 6_4, UpperCAmelCase__ : float = 0.1, UpperCAmelCase__ : float = 0.1, UpperCAmelCase__ : float = 0.1, UpperCAmelCase__ : float = 0.1, UpperCAmelCase__ : float = 0.1, UpperCAmelCase__ : int = 1_0_0, UpperCAmelCase__ : float = 0.02, UpperCAmelCase__ : Any=True, **UpperCAmelCase__ : List[str], ): # time series specific configuration __lowercase = prediction_length __lowercase = context_length or prediction_length __lowercase = distribution_output __lowercase = loss __lowercase = input_size __lowercase = num_time_features __lowercase = lags_sequence __lowercase = scaling __lowercase = num_dynamic_real_features __lowercase = num_static_real_features __lowercase = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(UpperCAmelCase__ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __lowercase = cardinality else: __lowercase = [0] if embedding_dimension and num_static_categorical_features > 0: if len(UpperCAmelCase__ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __lowercase = embedding_dimension else: __lowercase = [min(5_0, (cat + 1) // 2 ) for cat in self.cardinality] __lowercase = num_parallel_samples # Transformer architecture configuration __lowercase = input_size * len(UpperCAmelCase__ ) + self._number_of_features __lowercase = d_model __lowercase = encoder_attention_heads __lowercase = decoder_attention_heads __lowercase = encoder_ffn_dim __lowercase = decoder_ffn_dim __lowercase = encoder_layers __lowercase = decoder_layers __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = activation_function __lowercase = init_std __lowercase = use_cache super().__init__(is_encoder_decoder=UpperCAmelCase__, **UpperCAmelCase__ ) @property def _lowercase ( self : Optional[Any] ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
17
1
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase ): if not isinstance(_UpperCamelCase , _UpperCamelCase ): raise ValueError('iterations must be defined as integers' ) if not isinstance(_UpperCamelCase , _UpperCamelCase ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) __lowerCAmelCase : List[str] = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(_UpperCamelCase ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
182
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCamelCase__ = """hf-internal-testing/tiny-random-bert""" lowerCamelCase__ = os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") lowerCamelCase__ = """9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): __lowerCAmelCase : Union[str, Any] = cached_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(_SCREAMING_SNAKE_CASE ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'refs' , 'main' ) ) as f: __lowerCAmelCase : List[Any] = f.read() self.assertEqual(_SCREAMING_SNAKE_CASE , os.path.join(_SCREAMING_SNAKE_CASE , 'snapshots' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) self.assertTrue(os.path.isfile(_SCREAMING_SNAKE_CASE ) ) # File is cached at the same place the second time. __lowerCAmelCase : Union[str, Any] = cached_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Using a specific revision to test the full commit hash. __lowerCAmelCase : Union[str, Any] = cached_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , revision='9b8c223' ) self.assertEqual(_SCREAMING_SNAKE_CASE , os.path.join(_SCREAMING_SNAKE_CASE , 'snapshots' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'is not a valid model identifier' ): __lowerCAmelCase : Optional[Any] = cached_file('tiny-random-bert' , _SCREAMING_SNAKE_CASE ) with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'is not a valid git identifier' ): __lowerCAmelCase : str = cached_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , revision='aaaa' ) with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'does not appear to have a file named' ): __lowerCAmelCase : Optional[Any] = cached_file(_SCREAMING_SNAKE_CASE , 'conf' ) def __lowerCamelCase ( self ): with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'does not appear to have a file named' ): __lowerCAmelCase : Optional[int] = cached_file(_SCREAMING_SNAKE_CASE , 'conf' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'refs' , 'main' ) ) as f: __lowerCAmelCase : Tuple = f.read() self.assertTrue(os.path.isfile(os.path.join(_SCREAMING_SNAKE_CASE , '.no_exist' , _SCREAMING_SNAKE_CASE , 'conf' ) ) ) __lowerCAmelCase : List[Any] = cached_file(_SCREAMING_SNAKE_CASE , 'conf' , _raise_exceptions_for_missing_entries=_SCREAMING_SNAKE_CASE ) self.assertIsNone(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = cached_file(_SCREAMING_SNAKE_CASE , 'conf' , local_files_only=_SCREAMING_SNAKE_CASE , _raise_exceptions_for_missing_entries=_SCREAMING_SNAKE_CASE ) self.assertIsNone(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = mock.Mock() __lowerCAmelCase : Tuple = 5_00 __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : Dict = HTTPError __lowerCAmelCase : str = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=_SCREAMING_SNAKE_CASE ) as mock_head: __lowerCAmelCase : Optional[Any] = cached_file(_SCREAMING_SNAKE_CASE , 'conf' , _raise_exceptions_for_connection_errors=_SCREAMING_SNAKE_CASE ) self.assertIsNone(_SCREAMING_SNAKE_CASE ) # This check we did call the fake head request mock_head.assert_called() def __lowerCamelCase ( self ): self.assertTrue(has_file('hf-internal-testing/tiny-bert-pt-only' , _SCREAMING_SNAKE_CASE ) ) self.assertFalse(has_file('hf-internal-testing/tiny-bert-pt-only' , _SCREAMING_SNAKE_CASE ) ) self.assertFalse(has_file('hf-internal-testing/tiny-bert-pt-only' , _SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('bert-base-cased' , 'ahah.txt' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'is not a valid model identifier' ): get_file_from_repo('bert-base-case' , _SCREAMING_SNAKE_CASE ) # The function raises if the revision does not exist. with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , 'is not a valid git identifier' ): get_file_from_repo('bert-base-cased' , _SCREAMING_SNAKE_CASE , revision='ahaha' ) __lowerCAmelCase : Union[str, Any] = get_file_from_repo('bert-base-cased' , _SCREAMING_SNAKE_CASE ) # The name is the cached name which is not very easy to test, so instead we load the content. __lowerCAmelCase : List[Any] = json.loads(open(_SCREAMING_SNAKE_CASE , 'r' ).read() ) self.assertEqual(config['hidden_size'] , 7_68 ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCAmelCase : str = Path(_SCREAMING_SNAKE_CASE ) / 'a.txt' filename.touch() self.assertEqual(get_file_from_repo(_SCREAMING_SNAKE_CASE , 'a.txt' ) , str(_SCREAMING_SNAKE_CASE ) ) self.assertIsNone(get_file_from_repo(_SCREAMING_SNAKE_CASE , 'b.txt' ) )
182
1
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( __lowerCamelCase , unittest.TestCase ): _a = DiTPipeline _a = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS _a = PipelineTesterMixin.required_optional_params - { """latents""", """num_images_per_prompt""", """callback""", """callback_steps""", } _a = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS _a = False def A__ ( self ) -> int: '''simple docstring''' torch.manual_seed(0 ) _lowercase =TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=lowerCAmelCase , activation_fn='gelu-approximate' , num_embeds_ada_norm=1_000 , norm_type='ada_norm_zero' , norm_elementwise_affine=lowerCAmelCase , ) _lowercase =AutoencoderKL() _lowercase =DDIMScheduler() _lowercase ={'transformer': transformer.eval(), 'vae': vae.eval(), 'scheduler': scheduler} return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ) -> Any: '''simple docstring''' if str(lowerCAmelCase ).startswith('mps' ): _lowercase =torch.manual_seed(lowerCAmelCase ) else: _lowercase =torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) _lowercase ={ 'class_labels': [1], 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def A__ ( self ) -> str: '''simple docstring''' _lowercase ='cpu' _lowercase =self.get_dummy_components() _lowercase =self.pipeline_class(**lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) _lowercase =self.get_dummy_inputs(lowerCAmelCase ) _lowercase =pipe(**lowerCAmelCase ).images _lowercase =image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) _lowercase =np.array([0.2946, 0.6601, 0.4329, 0.3296, 0.4144, 0.5319, 0.7273, 0.5013, 0.4457] ) _lowercase =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase , 1e-3 ) def A__ ( self ) -> int: '''simple docstring''' self._test_inference_batch_single_identical(relax_max_difference=lowerCAmelCase , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def A__ ( self ) -> Union[str, Any]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class __lowerCAmelCase ( unittest.TestCase ): def A__ ( self ) -> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self ) -> int: '''simple docstring''' _lowercase =torch.manual_seed(0 ) _lowercase =DiTPipeline.from_pretrained('facebook/DiT-XL-2-256' ) pipe.to('cuda' ) _lowercase =['vase', 'umbrella', 'white shark', 'white wolf'] _lowercase =pipe.get_label_ids(lowerCAmelCase ) _lowercase =pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=40 , output_type='np' ).images for word, image in zip(lowerCAmelCase , lowerCAmelCase ): _lowercase =load_numpy( F'''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-2 def A__ ( self ) -> List[str]: '''simple docstring''' _lowercase =DiTPipeline.from_pretrained('facebook/DiT-XL-2-512' ) _lowercase =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to('cuda' ) _lowercase =['vase', 'umbrella'] _lowercase =pipe.get_label_ids(lowerCAmelCase ) _lowercase =torch.manual_seed(0 ) _lowercase =pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type='np' ).images for word, image in zip(lowerCAmelCase , lowerCAmelCase ): _lowercase =load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' F'''/dit/{word}_512.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-1
205
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging UpperCAmelCase = logging.get_logger(__name__) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = r'\w+[.]\d+' lowercase = re.findall(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for pat in pats: lowercase = key.replace(__SCREAMING_SNAKE_CASE , '_'.join(pat.split('.' ) ) ) return key def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): lowercase = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: lowercase = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: lowercase = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer lowercase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: lowercase = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer lowercase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": lowercase = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight lowercase = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias lowercase = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=42 ): # Step 1: Convert pytorch tensor to numpy lowercase = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params lowercase = flax_model.init_weights(PRNGKey(__SCREAMING_SNAKE_CASE ) ) lowercase = flatten_dict(__SCREAMING_SNAKE_CASE ) lowercase = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowercase = rename_key(__SCREAMING_SNAKE_CASE ) lowercase = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters lowercase , lowercase = rename_key_and_reshape_tensor(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'''PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ''' F'''{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) # also add unexpected weight so that warning is thrown lowercase = jnp.asarray(__SCREAMING_SNAKE_CASE ) return unflatten_dict(__SCREAMING_SNAKE_CASE )
195
0
import numpy as np def __snake_case ( _UpperCAmelCase ): return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
131
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __snake_case :Tuple = { '''configuration_bloom''': ['''BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BloomConfig''', '''BloomOnnxConfig'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :List[Any] = ['''BloomTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :Any = [ '''BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BloomForCausalLM''', '''BloomModel''', '''BloomPreTrainedModel''', '''BloomForSequenceClassification''', '''BloomForTokenClassification''', '''BloomForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys __snake_case :Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
131
1
'''simple docstring''' import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ ( unittest.TestCase ): def lowercase ( self : Tuple ): _UpperCAmelCase = "ylacombe/bark-small" _UpperCAmelCase = tempfile.mkdtemp() _UpperCAmelCase = "en_speaker_1" _UpperCAmelCase = "This is a test string" _UpperCAmelCase = "speaker_embeddings_path.json" _UpperCAmelCase = "speaker_embeddings" def lowercase ( self : Optional[Any] , **snake_case_ : List[Any] ): return AutoTokenizer.from_pretrained(self.checkpoint , **_UpperCamelCase ) def lowercase ( self : Union[str, Any] ): shutil.rmtree(self.tmpdirname ) def lowercase ( self : Optional[Any] ): _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BarkProcessor(tokenizer=_UpperCamelCase ) processor.save_pretrained(self.tmpdirname ) _UpperCAmelCase = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def lowercase ( self : List[Any] ): _UpperCAmelCase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) _UpperCAmelCase = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _UpperCAmelCase = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def lowercase ( self : int ): _UpperCAmelCase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) _UpperCAmelCase = 3_5 _UpperCAmelCase = 2 _UpperCAmelCase = 8 _UpperCAmelCase = { "semantic_prompt": np.ones(_UpperCamelCase ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset _UpperCAmelCase = processor(text=self.input_string , voice_preset=_UpperCamelCase ) _UpperCAmelCase = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(_UpperCamelCase , np.array([] ) ).tolist() ) # test loading voice preset from npz file _UpperCAmelCase = os.path.join(self.tmpdirname , "file.npz" ) np.savez(_UpperCamelCase , **_UpperCamelCase ) _UpperCAmelCase = processor(text=self.input_string , voice_preset=_UpperCamelCase ) _UpperCAmelCase = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(_UpperCamelCase , np.array([] ) ).tolist() ) # test loading voice preset from the hub _UpperCAmelCase = processor(text=self.input_string , voice_preset=self.voice_preset ) def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BarkProcessor(tokenizer=_UpperCamelCase ) _UpperCAmelCase = processor(text=self.input_string ) _UpperCAmelCase = tokenizer( self.input_string , padding="max_length" , max_length=2_5_6 , add_special_tokens=_UpperCamelCase , return_attention_mask=_UpperCamelCase , return_token_type_ids=_UpperCamelCase , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
22
'''simple docstring''' class a__ : def __init__( self , _UpperCamelCase ): """simple docstring""" _lowercase : Tuple = n _lowercase : Any = [None] * self.n _lowercase : Tuple = 0 # index of the first element _lowercase : Union[str, Any] = 0 _lowercase : str = 0 def __len__( self ): """simple docstring""" return self.size def _lowerCamelCase ( self ): """simple docstring""" return self.size == 0 def _lowerCamelCase ( self ): """simple docstring""" return False if self.is_empty() else self.array[self.front] def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" if self.size >= self.n: raise Exception("QUEUE IS FULL" ) _lowercase : Optional[int] = data _lowercase : Dict = (self.rear + 1) % self.n self.size += 1 return self def _lowerCamelCase ( self ): """simple docstring""" if self.size == 0: raise Exception("UNDERFLOW" ) _lowercase : Optional[Any] = self.array[self.front] _lowercase : List[Any] = None _lowercase : int = (self.front + 1) % self.n self.size -= 1 return temp
250
0
"""simple docstring""" def a__ ( __lowercase , __lowercase ) -> int: return 1 if input_a == input_a else 0 def a__ ( ) -> None: assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
163
"""simple docstring""" from collections.abc import Generator from math import sin def a__ ( __lowercase ) -> bytes: if len(__lowercase ) != 32: raise ValueError("Input must be of length 32" ) _A = B"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def a__ ( __lowercase ) -> bytes: if i < 0: raise ValueError("Input must be non-negative" ) _A = format(__lowercase , "08x" )[-8:] _A = B"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def a__ ( __lowercase ) -> bytes: _A = B"" for char in message: bit_string += format(__lowercase , "08b" ).encode("utf-8" ) _A = format(len(__lowercase ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__lowercase ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def a__ ( __lowercase ) -> Generator[list[int], None, None]: if len(__lowercase ) % 512 != 0: raise ValueError("Input must have length that's a multiple of 512" ) for pos in range(0 , len(__lowercase ) , 512 ): _A = bit_string[pos : pos + 512] _A = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def a__ ( __lowercase ) -> int: if i < 0: raise ValueError("Input must be non-negative" ) _A = format(__lowercase , "032b" ) _A = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(__lowercase , 2 ) def a__ ( __lowercase , __lowercase ) -> int: return (a + b) % 2**32 def a__ ( __lowercase , __lowercase ) -> int: if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def a__ ( __lowercase ) -> bytes: _A = preprocess(__lowercase ) _A = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states _A = 0x67_452_301 _A = 0xef_cda_b89 _A = 0x98_bad_cfe _A = 0x10_325_476 _A = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__lowercase ): _A = aa _A = ba _A = ca _A = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f _A = d ^ (b & (c ^ d)) _A = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f _A = c ^ (d & (b ^ c)) _A = (5 * i + 1) % 16 elif i <= 47: _A = b ^ c ^ d _A = (3 * i + 5) % 16 else: _A = c ^ (b | not_aa(__lowercase )) _A = (7 * i) % 16 _A = (f + a + added_consts[i] + block_words[g]) % 2**32 _A = d _A = c _A = b _A = sum_aa(__lowercase , left_rotate_aa(__lowercase , shift_amounts[i] ) ) # Add hashed chunk to running total _A = sum_aa(__lowercase , __lowercase ) _A = sum_aa(__lowercase , __lowercase ) _A = sum_aa(__lowercase , __lowercase ) _A = sum_aa(__lowercase , __lowercase ) _A = reformat_hex(__lowercase ) + reformat_hex(__lowercase ) + reformat_hex(__lowercase ) + reformat_hex(__lowercase ) return digest if __name__ == "__main__": import doctest doctest.testmod()
163
1
"""simple docstring""" import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCamelCase_ = '''\ ''' lowerCamelCase_ = ''' Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity ''' lowerCamelCase_ = ''' Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to \'cuda\' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"] >>> results = perplexity.compute(model_id=\'gpt2\', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 78.22 >>> print(round(results["perplexities"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = datasets.load_dataset("wikitext", ... "wikitext-2-raw-v1", ... split="test")["text"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=\'\'] >>> results = perplexity.compute(model_id=\'gpt2\', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 60.35 >>> print(round(results["perplexities"][0], 2)) 81.12 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase_ (datasets.Metric ): def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "input_texts": datasets.Value("string" ), } ) , reference_urls=["https://huggingface.co/docs/transformers/perplexity"] , ) def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : int = 16 , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : List[str]=None ) -> Union[str, Any]: if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCAmelCase_ : str = "cuda" else: UpperCAmelCase_ : Dict = "cuda" if torch.cuda.is_available() else "cpu" UpperCAmelCase_ : str = AutoModelForCausalLM.from_pretrained(lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = model.to(lowerCAmelCase_ ) UpperCAmelCase_ : Any = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCAmelCase_ : Tuple = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(lowerCAmelCase_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"pad_token": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCAmelCase_ : Tuple = model.config.max_length - 1 else: UpperCAmelCase_ : int = model.config.max_length UpperCAmelCase_ : Union[str, Any] = tokenizer( lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors="pt" , return_attention_mask=lowerCAmelCase_ , ).to(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = encodings["input_ids"] UpperCAmelCase_ : List[Any] = encodings["attention_mask"] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : int = CrossEntropyLoss(reduction="none" ) for start_index in logging.tqdm(range(0 , len(lowerCAmelCase_ ) , lowerCAmelCase_ ) ): UpperCAmelCase_ : Optional[Any] = min(start_index + batch_size , len(lowerCAmelCase_ ) ) UpperCAmelCase_ : int = encoded_texts[start_index:end_index] UpperCAmelCase_ : Union[str, Any] = attn_masks[start_index:end_index] if add_start_token: UpperCAmelCase_ : str = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCAmelCase_ : Tuple = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(lowerCAmelCase_ ), attn_mask] , dim=1 ) UpperCAmelCase_ : Tuple = encoded_batch with torch.no_grad(): UpperCAmelCase_ : Optional[int] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).logits UpperCAmelCase_ : Union[str, Any] = out_logits[..., :-1, :].contiguous() UpperCAmelCase_ : List[Any] = labels[..., 1:].contiguous() UpperCAmelCase_ : Tuple = attn_mask[..., 1:].contiguous() UpperCAmelCase_ : List[Any] = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , lowerCAmelCase_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(lowerCAmelCase_ )}
268
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase_ (__A ): def __init__( self : Optional[int] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict=768 ) -> List[Any]: super().__init__(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = proj_size UpperCAmelCase_ : Optional[Any] = CLIPVisionModel(lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = PaintByExampleMapper(lowerCAmelCase_ ) UpperCAmelCase_ : str = nn.LayerNorm(config.hidden_size ) UpperCAmelCase_ : List[Any] = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling UpperCAmelCase_ : Optional[int] = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict=False ) -> Union[str, Any]: UpperCAmelCase_ : Optional[int] = self.model(pixel_values=lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = clip_output.pooler_output UpperCAmelCase_ : List[Any] = self.mapper(latent_states[:, None] ) UpperCAmelCase_ : List[str] = self.final_layer_norm(lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = self.proj_out(lowerCAmelCase_ ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class UpperCamelCase_ (nn.Module ): def __init__( self : Dict , lowerCAmelCase_ : Union[str, Any] ) -> Tuple: super().__init__() UpperCAmelCase_ : List[Any] = (config.num_hidden_layers + 1) // 5 UpperCAmelCase_ : Optional[Any] = config.hidden_size UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ : Union[str, Any] = nn.ModuleList( [ BasicTransformerBlock(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , activation_fn="gelu" , attention_bias=lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ) ] ) def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase_ : List[str] ) -> str: for block in self.blocks: UpperCAmelCase_ : int = block(lowerCAmelCase_ ) return hidden_states
268
1
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __A : Optional[Any] = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __A : Optional[Any] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __A : int = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> tuple[str, float]: '''simple docstring''' lowerCAmelCase : Optional[Any] = len([g for position, g in enumerate(_UpperCAmelCase ) if g == main_target[position]] ) return (item, float(_UpperCAmelCase )) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> tuple[str, str]: '''simple docstring''' lowerCAmelCase : List[Any] = random.randint(0, len(_UpperCAmelCase ) - 1 ) lowerCAmelCase : str = parent_a[:random_slice] + parent_a[random_slice:] lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> str: '''simple docstring''' lowerCAmelCase : Optional[int] = list(_UpperCAmelCase ) if random.uniform(0, 1 ) < MUTATION_PROBABILITY: lowerCAmelCase : List[Any] = random.choice(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, ) -> list[str]: '''simple docstring''' lowerCAmelCase : Optional[int] = [] # Generate more children proportionally to the fitness score. lowerCAmelCase : Optional[Any] = int(parent_a[1] * 100 ) + 1 lowerCAmelCase : Optional[int] = 10 if child_n >= 10 else child_n for _ in range(_UpperCAmelCase ): lowerCAmelCase : Dict = population_score[random.randint(0, _UpperCAmelCase )][0] lowerCAmelCase : str = crossover(parent_a[0], _UpperCAmelCase ) # Append new string to the population list. pop.append(mutate(_UpperCAmelCase, _UpperCAmelCase ) ) pop.append(mutate(_UpperCAmelCase, _UpperCAmelCase ) ) return pop def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase = True ) -> tuple[int, int, str]: '''simple docstring''' if N_POPULATION < N_SELECTED: lowerCAmelCase : int = f"{N_POPULATION} must be bigger than {N_SELECTED}" raise ValueError(_UpperCAmelCase ) # Verify that the target contains no genes besides the ones inside genes variable. lowerCAmelCase : List[Any] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: lowerCAmelCase : Optional[int] = f"{not_in_genes_list} is not in genes list, evolution cannot converge" raise ValueError(_UpperCAmelCase ) # Generate random starting population. lowerCAmelCase : Any = [] for _ in range(_UpperCAmelCase ): population.append(''.join([random.choice(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) )] ) ) # Just some logs to know what the algorithms is doing. lowerCAmelCase : Optional[Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_UpperCAmelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. lowerCAmelCase : str = [evaluate(_UpperCAmelCase, _UpperCAmelCase ) for item in population] # Check if there is a matching evolution. lowerCAmelCase : Union[str, Any] = sorted(_UpperCAmelCase, key=lambda _UpperCAmelCase : x[1], reverse=_UpperCAmelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"\nGeneration: {generation}" f"\nTotal Population:{total_population}" f"\nBest score: {population_score[0][1]}" f"\nBest string: {population_score[0][0]}" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. lowerCAmelCase : Dict = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_UpperCAmelCase ) # Normalize population score to be between 0 and 1. lowerCAmelCase : List[Any] = [ (item, score / len(_UpperCAmelCase )) for item, score in population_score ] # This is selection for i in range(_UpperCAmelCase ): population.extend(select(population_score[int(_UpperCAmelCase )], _UpperCAmelCase, _UpperCAmelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_UpperCAmelCase ) > N_POPULATION: break if __name__ == "__main__": __A : int = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __A : Any = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __A : Union[str, Any] = basic(target_str, genes_list) print( F'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
363
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A : Union[str, Any] = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
323
0
'''simple docstring''' def __lowerCAmelCase ( UpperCamelCase__ ) -> List[Any]: if len(_lowerCAmelCase ) <= 1: return lst __lowerCamelCase = 1 while i < len(_lowerCAmelCase ): if lst[i - 1] <= lst[i]: i += 1 else: __lowerCamelCase , __lowerCamelCase = lst[i], lst[i - 1] i -= 1 if i == 0: __lowerCamelCase = 1 return lst if __name__ == "__main__": __UpperCAmelCase =input("Enter numbers separated by a comma:\n").strip() __UpperCAmelCase =[int(item) for item in user_input.split(",")] print(gnome_sort(unsorted))
67
"""simple docstring""" import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def lowercase (_lowerCAmelCase , _lowerCAmelCase="shi-labs/oneformer_demo" ): with open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) as f: __lowerCAmelCase = json.load(_lowerCAmelCase ) __lowerCAmelCase = {} __lowerCAmelCase = [] __lowerCAmelCase = [] for key, info in class_info.items(): __lowerCAmelCase = info["""name"""] class_names.append(info["""name"""] ) if info["isthing"]: thing_ids.append(int(_lowerCAmelCase ) ) __lowerCAmelCase = thing_ids __lowerCAmelCase = class_names return metadata class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=30 , snake_case_=400 , snake_case_=None , snake_case_=True , snake_case_=True , snake_case_=[0.5, 0.5, 0.5] , snake_case_=[0.5, 0.5, 0.5] , snake_case_=10 , snake_case_=False , snake_case_=255 , snake_case_="shi-labs/oneformer_demo" , snake_case_="ade20k_panoptic.json" , snake_case_=10 , ) -> Union[str, Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = {"""shortest_edge""": 32, """longest_edge""": 1_333} if size is None else size __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean __lowerCAmelCase = image_std __lowerCAmelCase = class_info_file __lowerCAmelCase = prepare_metadata(snake_case_ , snake_case_ ) __lowerCAmelCase = num_text __lowerCAmelCase = repo_path # for the post_process_functions __lowerCAmelCase = 2 __lowerCAmelCase = 10 __lowerCAmelCase = 10 __lowerCAmelCase = 3 __lowerCAmelCase = 4 __lowerCAmelCase = num_labels __lowerCAmelCase = do_reduce_labels __lowerCAmelCase = ignore_index def A__ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def A__ ( self , snake_case_ , snake_case_=False ) -> Dict: if not batched: __lowerCAmelCase = image_inputs[0] if isinstance(snake_case_ , Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image.size else: __lowerCAmelCase , __lowerCAmelCase = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase = int(self.size["""shortest_edge"""] * h / w ) __lowerCAmelCase = self.size["""shortest_edge"""] elif w > h: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = int(self.size["""shortest_edge"""] * w / h ) else: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = self.size["""shortest_edge"""] else: __lowerCAmelCase = [] for image in image_inputs: __lowerCAmelCase , __lowerCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[0] )[0] __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[1] )[1] return expected_height, expected_width def A__ ( self ) -> Tuple: return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class lowerCAmelCase_ ( A__ , unittest.TestCase ): '''simple docstring''' _snake_case = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string _snake_case = image_processing_class def A__ ( self ) -> str: __lowerCAmelCase = OneFormerImageProcessorTester(self ) @property def A__ ( self ) -> Dict: return self.image_processing_tester.prepare_image_processor_dict() def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case_ , """image_std""" ) ) self.assertTrue(hasattr(snake_case_ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case_ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case_ , """size""" ) ) self.assertTrue(hasattr(snake_case_ , """ignore_index""" ) ) self.assertTrue(hasattr(snake_case_ , """class_info_file""" ) ) self.assertTrue(hasattr(snake_case_ , """num_text""" ) ) self.assertTrue(hasattr(snake_case_ , """repo_path""" ) ) self.assertTrue(hasattr(snake_case_ , """metadata""" ) ) self.assertTrue(hasattr(snake_case_ , """do_reduce_labels""" ) ) def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Union[str, Any]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> List[str]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> Tuple: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self , snake_case_=False , snake_case_=False , snake_case_="np" ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # prepare image and target __lowerCAmelCase = self.image_processing_tester.num_labels __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) if with_segmentation_maps: __lowerCAmelCase = num_labels if is_instance_map: __lowerCAmelCase = list(range(snake_case_ ) ) * 2 __lowerCAmelCase = dict(enumerate(snake_case_ ) ) __lowerCAmelCase = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": __lowerCAmelCase = [Image.fromarray(snake_case_ ) for annotation in annotations] __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , snake_case_ , return_tensors="""pt""" , instance_id_to_semantic_id=snake_case_ , pad_and_return_pixel_mask=snake_case_ , ) return inputs def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Optional[Any]: def common(snake_case_=False , snake_case_=None ): __lowerCAmelCase = self.comm_get_image_processor_inputs( with_segmentation_maps=snake_case_ , is_instance_map=snake_case_ , segmentation_type=snake_case_ ) __lowerCAmelCase = inputs["""mask_labels"""] __lowerCAmelCase = inputs["""class_labels"""] __lowerCAmelCase = inputs["""pixel_values"""] __lowerCAmelCase = inputs["""text_inputs"""] # check the batch_size for mask_label, class_label, text_input in zip(snake_case_ , snake_case_ , snake_case_ ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.num_text ) common() common(is_instance_map=snake_case_ ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase = np.zeros((20, 50) ) __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = binary_mask_to_rle(snake_case_ ) self.assertEqual(len(snake_case_ ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) __lowerCAmelCase = [(1, 4) for i in range(self.image_processing_tester.batch_size )] __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ , target_sizes=snake_case_ ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_instance_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_panoptic_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
301
0
"""simple docstring""" __snake_case : Any = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] __snake_case : Union[str, Any] = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] __snake_case : int = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] __snake_case : Dict = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] __snake_case : Dict = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] __snake_case : Any = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] __snake_case : Tuple = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] __snake_case : str = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
58
"""simple docstring""" from ....utils import logging __snake_case : Optional[Any] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: str , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Optional[Any]=None , _SCREAMING_SNAKE_CASE: int=2048) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Any = config.__dict__ __lowerCAmelCase : Dict = modal_hidden_size if num_labels: __lowerCAmelCase : Optional[int] = num_labels
58
1
'''simple docstring''' def _lowerCamelCase ( lowercase : int , lowercase : int ) -> str: return base * power(_a , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('Raise base to the power of exponent using recursion...') lowerCAmelCase_ : Any = int(input('Enter the base: ').strip()) lowerCAmelCase_ : Dict = int(input('Enter the exponent: ').strip()) lowerCAmelCase_ : Union[str, Any] = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents lowerCAmelCase_ : Optional[Any] = 1 / result print(f"""{base} to the power of {exponent} is {result}""")
63
'''simple docstring''' from ..utils import DummyObject, requires_backends class _a ( metaclass=__a ): __a : int = ["""flax""", """transformers"""] def __init__( self : Optional[Any] , *lowercase : str , **lowercase : List[Any] ): '''simple docstring''' requires_backends(self , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Union[str, Any] , *lowercase : List[Any] , **lowercase : List[str] ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Union[str, Any] , *lowercase : Tuple , **lowercase : int ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) class _a ( metaclass=__a ): __a : int = ["""flax""", """transformers"""] def __init__( self : int , *lowercase : Tuple , **lowercase : Dict ): '''simple docstring''' requires_backends(self , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : List[str] , *lowercase : Optional[int] , **lowercase : List[Any] ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Dict , *lowercase : Union[str, Any] , **lowercase : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) class _a ( metaclass=__a ): __a : int = ["""flax""", """transformers"""] def __init__( self : Optional[int] , *lowercase : Union[str, Any] , **lowercase : Any ): '''simple docstring''' requires_backends(self , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Union[str, Any] , *lowercase : Tuple , **lowercase : Any ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Any , *lowercase : Dict , **lowercase : Optional[Any] ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) class _a ( metaclass=__a ): __a : Any = ["""flax""", """transformers"""] def __init__( self : Any , *lowercase : Optional[Any] , **lowercase : Optional[int] ): '''simple docstring''' requires_backends(self , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Dict , *lowercase : Optional[Any] , **lowercase : Optional[Any] ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] ) @classmethod def A ( cls : Union[str, Any] , *lowercase : str , **lowercase : Any ): '''simple docstring''' requires_backends(cls , ['''flax''', '''transformers'''] )
34
0
'''simple docstring''' import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json _UpperCAmelCase : str = """sshleifer/mar_enro_6_3_student""" class a__ ( __A ): """simple docstring""" def _snake_case (self ): super().setUp() __lowerCAmelCase = cached_path( '''https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz''' , extract_compressed_file=__lowercase , ) __lowerCAmelCase = F"""{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k""" @slow @require_torch_gpu def _snake_case (self ): MarianMTModel.from_pretrained(__lowercase ) @slow @require_torch_gpu def _snake_case (self ): __lowerCAmelCase = { '''$MAX_LEN''': 64, '''$BS''': 64, '''$GAS''': 1, '''$ENRO_DIR''': self.data_dir, '''facebook/mbart-large-cc25''': MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", '''--learning_rate=3e-5''': '''--learning_rate 3e-4''', '''--num_train_epochs 6''': '''--num_train_epochs 1''', } # Clean up bash script __lowerCAmelCase = (self.test_file_dir / '''train_mbart_cc25_enro.sh''').open().read().split('''finetune.py''' )[1].strip() __lowerCAmelCase = bash_script.replace('''\\\n''' , '''''' ).strip().replace('''"$@"''' , '''''' ) for k, v in env_vars_to_replace.items(): __lowerCAmelCase = bash_script.replace(__lowercase , str(__lowercase ) ) __lowerCAmelCase = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") __lowerCAmelCase = F""" --output_dir {output_dir} --tokenizer_name Helsinki-NLP/opus-mt-en-ro --sortish_sampler --do_predict --gpus 1 --freeze_encoder --n_train 40000 --n_val 500 --n_test 500 --fp16_opt_level O1 --num_sanity_val_steps 0 --eval_beams 2 """.split() # XXX: args.gpus > 1 : handle multi_gpu in the future __lowerCAmelCase = ['''finetune.py'''] + bash_script.split() + args with patch.object(__lowercase , '''argv''' , __lowercase ): __lowerCAmelCase = argparse.ArgumentParser() __lowerCAmelCase = pl.Trainer.add_argparse_args(__lowercase ) __lowerCAmelCase = SummarizationModule.add_model_specific_args(__lowercase , os.getcwd() ) __lowerCAmelCase = parser.parse_args() __lowerCAmelCase = main(__lowercase ) # Check metrics __lowerCAmelCase = load_json(model.metrics_save_path ) __lowerCAmelCase = metrics['''val'''][0] __lowerCAmelCase = metrics['''val'''][-1] self.assertEqual(len(metrics['''val'''] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , __lowercase ) self.assertGreater(last_step_stats['''val_avg_gen_time'''] , 0.0_1 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats['''val_avg_gen_time'''] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats['''val_avg_bleu'''] - first_step_stats['''val_avg_bleu'''] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats['''val_avg_bleu'''] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics['''val'''][-1]['''val_avg_bleu'''] - metrics['''test'''][-1]['''test_avg_bleu'''] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict __lowerCAmelCase = os.listdir(__lowercase ) __lowerCAmelCase = [x for x in contents if x.endswith('''.ckpt''' )][0] __lowerCAmelCase = os.path.join(args.output_dir , __lowercase ) __lowerCAmelCase = torch.load(__lowercase , map_location='''cpu''' ) __lowerCAmelCase = '''model.model.decoder.layers.0.encoder_attn_layer_norm.weight''' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: __lowerCAmelCase = {os.path.basename(__lowercase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['''test'''] ) == 1 class a__ ( __A ): """simple docstring""" @timeout_decorator.timeout(6_00 ) @slow @require_torch_gpu def _snake_case (self ): __lowerCAmelCase = F"""{self.test_file_dir_str}/test_data/wmt_en_ro""" __lowerCAmelCase = { '''--fp16_opt_level=O1''': '''''', '''$MAX_LEN''': 1_28, '''$BS''': 16, '''$GAS''': 1, '''$ENRO_DIR''': data_dir, '''$m''': '''sshleifer/student_marian_en_ro_6_1''', '''val_check_interval=0.25''': '''val_check_interval=1.0''', } # Clean up bash script __lowerCAmelCase = ( (self.test_file_dir / '''distil_marian_no_teacher.sh''').open().read().split('''distillation.py''' )[1].strip() ) __lowerCAmelCase = bash_script.replace('''\\\n''' , '''''' ).strip().replace('''"$@"''' , '''''' ) __lowerCAmelCase = bash_script.replace('''--fp16 ''' , ''' ''' ) for k, v in env_vars_to_replace.items(): __lowerCAmelCase = bash_script.replace(__lowercase , str(__lowercase ) ) __lowerCAmelCase = self.get_auto_remove_tmp_dir() __lowerCAmelCase = bash_script.replace('''--fp16''' , '''''' ) __lowerCAmelCase = 6 __lowerCAmelCase = ( ['''distillation.py'''] + bash_script.split() + [ F"""--output_dir={output_dir}""", '''--gpus=1''', '''--learning_rate=1e-3''', F"""--num_train_epochs={epochs}""", '''--warmup_steps=10''', '''--val_check_interval=1.0''', '''--do_predict''', ] ) with patch.object(__lowercase , '''argv''' , __lowercase ): __lowerCAmelCase = argparse.ArgumentParser() __lowerCAmelCase = pl.Trainer.add_argparse_args(__lowercase ) __lowerCAmelCase = SummarizationDistiller.add_model_specific_args(__lowercase , os.getcwd() ) __lowerCAmelCase = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu __lowerCAmelCase = distill_main(__lowercase ) # Check metrics __lowerCAmelCase = load_json(model.metrics_save_path ) __lowerCAmelCase = metrics['''val'''][0] __lowerCAmelCase = metrics['''val'''][-1] assert len(metrics['''val'''] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.0_1 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , __lowercase ) # check lightning ckpt can be loaded and has a reasonable statedict __lowerCAmelCase = os.listdir(__lowercase ) __lowerCAmelCase = [x for x in contents if x.endswith('''.ckpt''' )][0] __lowerCAmelCase = os.path.join(args.output_dir , __lowercase ) __lowerCAmelCase = torch.load(__lowercase , map_location='''cpu''' ) __lowerCAmelCase = '''model.model.decoder.layers.0.encoder_attn_layer_norm.weight''' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: __lowerCAmelCase = {os.path.basename(__lowercase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['''test'''] ) == 1
9
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self ): __lowerCAmelCase = tempfile.mkdtemp() # fmt: off __lowerCAmelCase = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on __lowerCAmelCase = dict(zip(__lowercase , range(len(__lowercase ) ) ) ) __lowerCAmelCase = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] __lowerCAmelCase = {'''unk_token''': '''<unk>'''} __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowercase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowercase ) ) __lowerCAmelCase = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], '''image_std''': [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } __lowerCAmelCase = os.path.join(self.tmpdirname , __lowercase ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(__lowercase , __lowercase ) def _snake_case (self , **__lowercase ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def _snake_case (self , **__lowercase ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **__lowercase ) def _snake_case (self , **__lowercase ): return CLIPImageProcessor.from_pretrained(self.tmpdirname , **__lowercase ) def _snake_case (self ): shutil.rmtree(self.tmpdirname ) def _snake_case (self ): __lowerCAmelCase = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] __lowerCAmelCase = [Image.fromarray(np.moveaxis(__lowercase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _snake_case (self ): __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = self.get_rust_tokenizer() __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) processor_slow.save_pretrained(self.tmpdirname ) __lowerCAmelCase = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=__lowercase ) __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) processor_fast.save_pretrained(self.tmpdirname ) __lowerCAmelCase = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowercase ) self.assertIsInstance(processor_fast.tokenizer , __lowercase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowercase ) self.assertIsInstance(processor_fast.image_processor , __lowercase ) def _snake_case (self ): __lowerCAmelCase = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase = self.get_image_processor(do_normalize=__lowercase , padding_value=1.0 ) __lowerCAmelCase = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__lowercase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowercase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowercase ) def _snake_case (self ): __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) __lowerCAmelCase = self.prepare_image_inputs() __lowerCAmelCase = image_processor(__lowercase , return_tensors='''np''' ) __lowerCAmelCase = processor(images=__lowercase , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _snake_case (self ): __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) __lowerCAmelCase = '''lower newer''' __lowerCAmelCase = processor(text=__lowercase ) __lowerCAmelCase = tokenizer(__lowercase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _snake_case (self ): __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) __lowerCAmelCase = '''lower newer''' __lowerCAmelCase = self.prepare_image_inputs() __lowerCAmelCase = processor(text=__lowercase , images=__lowercase ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(__lowercase ): processor() def _snake_case (self ): __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) __lowerCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase = processor.batch_decode(__lowercase ) __lowerCAmelCase = tokenizer.batch_decode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def _snake_case (self ): __lowerCAmelCase = self.get_image_processor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = CLIPProcessor(tokenizer=__lowercase , image_processor=__lowercase ) __lowerCAmelCase = '''lower newer''' __lowerCAmelCase = self.prepare_image_inputs() __lowerCAmelCase = processor(text=__lowercase , images=__lowercase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
9
1
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = """▁""" __lowercase = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} __lowercase = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } __lowercase = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } __lowercase = { """ernie-m-base""": 514, """ernie-m-large""": 514, } __lowercase = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class _A ( _a ): """simple docstring""" UpperCAmelCase : List[str] = ["input_ids"] UpperCAmelCase : Union[str, Any] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[Any] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : int = RESOURCE_FILES_NAMES def __init__( self : int , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : int=None , __UpperCAmelCase : Union[str, Any]=False , __UpperCAmelCase : Tuple="utf8" , __UpperCAmelCase : List[str]="[UNK]" , __UpperCAmelCase : Dict="[SEP]" , __UpperCAmelCase : Any="[PAD]" , __UpperCAmelCase : str="[CLS]" , __UpperCAmelCase : Optional[Any]="[MASK]" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Dict , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , vocab_file=__UpperCAmelCase , encoding=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) a : str = do_lower_case a : Dict = sentencepiece_model_ckpt a : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(__UpperCAmelCase) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: a : Any = self.load_vocab(filepath=__UpperCAmelCase) else: a : int = {self.sp_model.id_to_piece(__UpperCAmelCase): id for id in range(self.sp_model.get_piece_size())} a : int = {v: k for k, v in self.vocab.items()} def __snake_case ( self : Any , __UpperCAmelCase : List[Any]): if text is None: return None a : Any = self.tokenize(__UpperCAmelCase) a , a : str = "", [] for i, ch in enumerate(__UpperCAmelCase): if ch in self.SP_CHAR_MAPPING: a : List[Any] = self.SP_CHAR_MAPPING.get(__UpperCAmelCase) else: a : Union[str, Any] = unicodedata.normalize("NFKC" , __UpperCAmelCase) if self.is_whitespace(__UpperCAmelCase): continue normalized_text += ch char_mapping.extend([i] * len(__UpperCAmelCase)) a , a , a : Optional[int] = normalized_text, [], 0 if self.do_lower_case: a : List[str] = text.lower() for token in split_tokens: if token[:1] == "▁": a : int = token[1:] a : Tuple = text[offset:].index(__UpperCAmelCase) + offset a : List[str] = start + len(__UpperCAmelCase) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1)) a : List[Any] = end return token_mapping @property def __snake_case ( self : str): return len(self.vocab) def __snake_case ( self : Optional[Any]): return dict(self.vocab , **self.added_tokens_encoder) def __getstate__( self : int): a : str = self.__dict__.copy() a : Optional[Any] = None return state def __setstate__( self : str , __UpperCAmelCase : str): a : Tuple = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): a : Optional[int] = {} a : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.sentencepiece_model_ckpt) def __snake_case ( self : Dict , __UpperCAmelCase : List[Any]): return "".join((self.SP_CHAR_MAPPING.get(__UpperCAmelCase , __UpperCAmelCase) for c in text)) def __snake_case ( self : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : Optional[Any]=64 , __UpperCAmelCase : Any=0.1): if self.sp_model_kwargs.get("enable_sampling") is True: a : Union[str, Any] = True if self.sp_model_kwargs.get("alpha") is not None: a : Optional[int] = self.sp_model_kwargs.get("alpha") if self.sp_model_kwargs.get("nbest_size") is not None: a : Dict = self.sp_model_kwargs.get("nbest_size") if not enable_sampling: a : int = self.sp_model.EncodeAsPieces(__UpperCAmelCase) else: a : List[Any] = self.sp_model.SampleEncodeAsPieces(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) a : Optional[Any] = [] for pi, piece in enumerate(__UpperCAmelCase): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(__UpperCAmelCase) and pi != 0: new_pieces.append(__UpperCAmelCase) continue else: continue a : Any = 0 for i, chunk in enumerate(__UpperCAmelCase): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(__UpperCAmelCase) or self.is_punct(__UpperCAmelCase): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) new_pieces.append(__UpperCAmelCase) a : Optional[int] = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) a : int = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) a : Tuple = i if len(__UpperCAmelCase) > lst_i: new_pieces.append(piece[lst_i:]) return new_pieces def __snake_case ( self : List[str] , __UpperCAmelCase : Union[str, Any]): a : str = "".join(__UpperCAmelCase).replace(__UpperCAmelCase , " ").strip() return out_string def __snake_case ( self : Optional[int] , __UpperCAmelCase : Union[str, Any]): a : Tuple = self.convert_ids_to_tokens(__UpperCAmelCase) a : str = "".join(__UpperCAmelCase).replace(__UpperCAmelCase , " ").strip() return out_string def __snake_case ( self : Union[str, Any] , __UpperCAmelCase : Optional[int]): return self.vocab.get(__UpperCAmelCase , self.vocab.get(self.unk_token)) def __snake_case ( self : Any , __UpperCAmelCase : Tuple): return self.reverse_vocab.get(__UpperCAmelCase , self.unk_token) def __snake_case ( self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : Union[str, Any]=None): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a : int = [self.cls_token_id] a : Optional[int] = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def __snake_case ( self : List[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None): if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def __snake_case ( self : int , __UpperCAmelCase : Tuple , __UpperCAmelCase : str=None , __UpperCAmelCase : int=False): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model.") return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase)) + [1, 1] + ([0] * len(__UpperCAmelCase)) + [1] return [1] + ([0] * len(__UpperCAmelCase)) + [1] def __snake_case ( self : str , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None): # called when `add_special_tokens` is True, so align with `build_inputs_with_special_tokens` method if token_ids_a is None: # [CLS] X [SEP] return (len(__UpperCAmelCase) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(__UpperCAmelCase) + 1) + [1] * (len(__UpperCAmelCase) + 3) def __snake_case ( self : int , __UpperCAmelCase : Optional[int]): if "\u4e00" <= char <= "\u9fff": return True return False def __snake_case ( self : Tuple , __UpperCAmelCase : Union[str, Any]): if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def __snake_case ( self : int , __UpperCAmelCase : List[Any]): if char in ",;:.?!~,;:。?!《》【】": return True return False def __snake_case ( self : str , __UpperCAmelCase : Optional[Any]): if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(__UpperCAmelCase) == 1: a : Optional[Any] = unicodedata.category(__UpperCAmelCase) if cat == "Zs": return True return False def __snake_case ( self : Optional[int] , __UpperCAmelCase : Optional[Any]): a : Optional[int] = {} with io.open(__UpperCAmelCase , "r" , encoding="utf-8") as f: for index, line in enumerate(__UpperCAmelCase): a : Tuple = line.rstrip("\n") a : Any = int(__UpperCAmelCase) return token_to_idx def __snake_case ( self : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None): a : int = 0 if os.path.isdir(__UpperCAmelCase): a : Union[str, Any] = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) else: a : Tuple = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(__UpperCAmelCase , "w" , encoding="utf-8") as writer: for token, token_index in sorted(self.vocab.items() , key=lambda __UpperCAmelCase: kv[1]): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!") a : Union[str, Any] = token_index writer.write(token + "\n") index += 1 a : Union[str, Any] = os.path.join(__UpperCAmelCase , "sentencepiece.bpe.model") with open(__UpperCAmelCase , "wb") as fi: a : Optional[int] = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase) return (vocab_file,)
40
"""simple docstring""" import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class a ( __snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE : List[str] = MobileBertTokenizer SCREAMING_SNAKE_CASE : int = MobileBertTokenizerFast SCREAMING_SNAKE_CASE : Union[str, Any] = True SCREAMING_SNAKE_CASE : Optional[Any] = True SCREAMING_SNAKE_CASE : Dict = filter_non_english SCREAMING_SNAKE_CASE : str = """google/mobilebert-uncased""" def UpperCamelCase ( self : List[str] ) -> Dict: super().setUp() lowerCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) lowerCamelCase_ = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def UpperCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] ) -> int: lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = 'unwanted, running' return input_text, output_text def UpperCamelCase ( self : Dict ) -> Any: lowerCamelCase_ = self.tokenizer_class(self.vocab_file ) lowerCamelCase_ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(__SCREAMING_SNAKE_CASE , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__SCREAMING_SNAKE_CASE ) , [9, 6, 7, 12, 10, 11] ) def UpperCamelCase ( self : Dict ) -> Dict: if not self.test_rust_tokenizer: return lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # With lower casing lowerCamelCase_ = self.get_tokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = self.get_rust_tokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self : List[str] ) -> str: lowerCamelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def UpperCamelCase ( self : List[Any] ) -> Optional[int]: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def UpperCamelCase ( self : Any ) -> Any: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def UpperCamelCase ( self : Tuple ) -> Union[str, Any]: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def UpperCamelCase ( self : Tuple ) -> List[str]: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def UpperCamelCase ( self : Optional[int] ) -> Any: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self : Tuple ) -> Tuple: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self : List[str] ) -> List[Any]: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self : Tuple ) -> str: lowerCamelCase_ = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def UpperCamelCase ( self : List[str] ) -> Any: lowerCamelCase_ = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] lowerCamelCase_ = {} for i, token in enumerate(__SCREAMING_SNAKE_CASE ): lowerCamelCase_ = i lowerCamelCase_ = WordpieceTokenizer(vocab=__SCREAMING_SNAKE_CASE , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def UpperCamelCase ( self : List[Any] ) -> Any: self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def UpperCamelCase ( self : Union[str, Any] ) -> int: self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def UpperCamelCase ( self : str ) -> Optional[Any]: self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def UpperCamelCase ( self : int ) -> List[Any]: lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def UpperCamelCase ( self : Dict ) -> List[str]: lowerCamelCase_ = self.tokenizer_class.from_pretrained('google/mobilebert-uncased' ) lowerCamelCase_ = tokenizer.encode('sequence builders' , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer.encode('multi-sequence build' , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def UpperCamelCase ( self : Tuple ) -> Union[str, Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = F'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' lowerCamelCase_ = tokenizer_r.encode_plus( __SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , return_token_type_ids=__SCREAMING_SNAKE_CASE , return_offsets_mapping=__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE , ) lowerCamelCase_ = tokenizer_r.do_lower_case if hasattr(__SCREAMING_SNAKE_CASE , 'do_lower_case' ) else False lowerCamelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def UpperCamelCase ( self : Any ) -> List[Any]: lowerCamelCase_ = ['的', '人', '有'] lowerCamelCase_ = ''.join(__SCREAMING_SNAKE_CASE ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_p.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_r.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCamelCase_ = False lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_r.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_p.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) # it is expected that only the first Chinese character is not preceded by "##". lowerCamelCase_ = [ F'''##{token}''' if idx != 0 else token for idx, token in enumerate(__SCREAMING_SNAKE_CASE ) ] self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
183
0
def lowerCAmelCase( SCREAMING_SNAKE_CASE_ )-> list: """simple docstring""" def merge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(SCREAMING_SNAKE_CASE_ ) <= 1: return collection UpperCamelCase_ = len(SCREAMING_SNAKE_CASE_ ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE :List[str] = input("""Enter numbers separated by a comma:\n""").strip() SCREAMING_SNAKE_CASE :Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
357
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class __magic_name__ : def __init__( self , _lowercase , _lowercase=2 , _lowercase=32 , _lowercase=16 , _lowercase=3 , _lowercase=True , _lowercase=True , _lowercase=32 , _lowercase=4 , _lowercase=[0, 1, 2, 3] , _lowercase=4 , _lowercase=37 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.02 , _lowercase=3 , _lowercase=[1, 384, 24, 24] , _lowercase=True , _lowercase=None , )-> List[Any]: UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = image_size UpperCamelCase_ = patch_size UpperCamelCase_ = num_channels UpperCamelCase_ = is_training UpperCamelCase_ = use_labels UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = backbone_out_indices UpperCamelCase_ = num_attention_heads UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = initializer_range UpperCamelCase_ = num_labels UpperCamelCase_ = backbone_featmap_shape UpperCamelCase_ = scope UpperCamelCase_ = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) UpperCamelCase_ = (image_size // patch_size) ** 2 UpperCamelCase_ = num_patches + 1 def UpperCAmelCase_ ( self )-> List[Any]: UpperCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase_ = None if self.use_labels: UpperCamelCase_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase_ = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self )-> List[str]: UpperCamelCase_ = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [96, 192, 384, 768], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=_lowercase , backbone_featmap_shape=self.backbone_featmap_shape , ) def UpperCAmelCase_ ( self , _lowercase , _lowercase , _lowercase )-> Any: UpperCamelCase_ = DPTModel(config=_lowercase ) model.to(_lowercase ) model.eval() UpperCamelCase_ = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowercase , _lowercase , _lowercase )-> Union[str, Any]: UpperCamelCase_ = self.num_labels UpperCamelCase_ = DPTForDepthEstimation(_lowercase ) model.to(_lowercase ) model.eval() UpperCamelCase_ = model(_lowercase ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowercase , _lowercase , _lowercase )-> str: UpperCamelCase_ = self.num_labels UpperCamelCase_ = DPTForSemanticSegmentation(_lowercase ) model.to(_lowercase ) model.eval() UpperCamelCase_ = model(_lowercase , labels=_lowercase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self )-> Union[str, Any]: UpperCamelCase_ = self.prepare_config_and_inputs() UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = config_and_inputs UpperCamelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( snake_case , snake_case , unittest.TestCase ): UpperCamelCase_ :Dict = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () UpperCamelCase_ :Any = ( { """depth-estimation""": DPTForDepthEstimation, """feature-extraction""": DPTModel, """image-segmentation""": DPTForSemanticSegmentation, } if is_torch_available() else {} ) UpperCamelCase_ :List[Any] = False UpperCamelCase_ :Dict = False UpperCamelCase_ :Tuple = False def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ = DPTModelTester(self ) UpperCamelCase_ = ConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def UpperCAmelCase_ ( self )-> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason="DPT does not use inputs_embeds" ) def UpperCAmelCase_ ( self )-> Any: pass def UpperCAmelCase_ ( self )-> Dict: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ = model_class(_lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowercase , nn.Linear ) ) def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ = model_class(_lowercase ) UpperCamelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase_ = [*signature.parameters.keys()] UpperCamelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*_lowercase ) def UpperCAmelCase_ ( self )-> Dict: UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowercase ) def UpperCAmelCase_ ( self )-> Optional[Any]: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ = True if model_class in get_values(_lowercase ): continue UpperCamelCase_ = model_class(_lowercase ) model.to(_lowercase ) model.train() UpperCamelCase_ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) UpperCamelCase_ = model(**_lowercase ).loss loss.backward() def UpperCAmelCase_ ( self )-> Optional[int]: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ = False UpperCamelCase_ = True if model_class in get_values(_lowercase ) or not model_class.supports_gradient_checkpointing: continue UpperCamelCase_ = model_class(_lowercase ) model.to(_lowercase ) model.gradient_checkpointing_enable() model.train() UpperCamelCase_ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) UpperCamelCase_ = model(**_lowercase ).loss loss.backward() def UpperCAmelCase_ ( self )-> Dict: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ = _config_zero_init(_lowercase ) for model_class in self.all_model_classes: UpperCamelCase_ = model_class(config=_lowercase ) # Skip the check for the backbone UpperCamelCase_ = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": UpperCamelCase_ = [F"{name}.{key}" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCAmelCase_ ( self )-> int: pass @slow def UpperCAmelCase_ ( self )-> List[Any]: for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: UpperCamelCase_ = DPTModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def UpperCAmelCase_ ( self )-> List[str]: # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ = "add" with self.assertRaises(_lowercase ): UpperCamelCase_ = DPTForDepthEstimation(_lowercase ) def lowerCAmelCase( )-> Any: """simple docstring""" UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision @slow class __magic_name__ ( unittest.TestCase ): def UpperCAmelCase_ ( self )-> Any: UpperCamelCase_ = DPTImageProcessor.from_pretrained("Intel/dpt-hybrid-midas" ) UpperCamelCase_ = DPTForDepthEstimation.from_pretrained("Intel/dpt-hybrid-midas" ).to(_lowercase ) UpperCamelCase_ = prepare_img() UpperCamelCase_ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): UpperCamelCase_ = model(**_lowercase ) UpperCamelCase_ = outputs.predicted_depth # verify the predicted depth UpperCamelCase_ = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , _lowercase ) UpperCamelCase_ = torch.tensor( [[[5.6_437, 5.6_146, 5.6_511], [5.4_371, 5.5_649, 5.5_958], [5.5_215, 5.5_184, 5.5_293]]] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , _lowercase , atol=1e-4 ) )
60
0
class __a : def __init__( self : List[str] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[int]=None , SCREAMING_SNAKE_CASE : Union[str, Any]=None ): '''simple docstring''' UpperCamelCase__ : int = data UpperCamelCase__ : Union[str, Any] = previous UpperCamelCase__ : Union[str, Any] = next_node def __str__( self : Any ): '''simple docstring''' return F'{self.data}' def __lowercase ( self : Optional[int] ): '''simple docstring''' return self.data def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.next def __lowercase ( self : Optional[Any] ): '''simple docstring''' return self.previous class __a : def __init__( self : str , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' UpperCamelCase__ : str = head def __iter__( self : Union[str, Any] ): '''simple docstring''' return self def __lowercase ( self : str ): '''simple docstring''' if not self.current: raise StopIteration else: UpperCamelCase__ : int = self.current.get_data() UpperCamelCase__ : int = self.current.get_next() return value class __a : def __init__( self : Tuple ): '''simple docstring''' UpperCamelCase__ : int = None # First node in list UpperCamelCase__ : Optional[int] = None # Last node in list def __str__( self : Tuple ): '''simple docstring''' UpperCamelCase__ : str = self.head UpperCamelCase__ : Optional[Any] = [] while current is not None: nodes.append(current.get_data() ) UpperCamelCase__ : int = current.get_next() return " ".join(str(SCREAMING_SNAKE_CASE ) for node in nodes ) def __contains__( self : List[Any] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' UpperCamelCase__ : Any = self.head while current: if current.get_data() == value: return True UpperCamelCase__ : Dict = current.get_next() return False def __iter__( self : Optional[Any] ): '''simple docstring''' return LinkedListIterator(self.head ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' if self.head: return self.head.get_data() return None def __lowercase ( self : List[str] ): '''simple docstring''' if self.tail: return self.tail.get_data() return None def __lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE : Node ): '''simple docstring''' if self.head is None: UpperCamelCase__ : Union[str, Any] = node UpperCamelCase__ : Union[str, Any] = node else: self.insert_before_node(self.head , SCREAMING_SNAKE_CASE ) def __lowercase ( self : Tuple , SCREAMING_SNAKE_CASE : Node ): '''simple docstring''' if self.head is None: self.set_head(SCREAMING_SNAKE_CASE ) else: self.insert_after_node(self.tail , SCREAMING_SNAKE_CASE ) def __lowercase ( self : List[str] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' UpperCamelCase__ : Optional[int] = Node(SCREAMING_SNAKE_CASE ) if self.head is None: self.set_head(SCREAMING_SNAKE_CASE ) else: self.set_tail(SCREAMING_SNAKE_CASE ) def __lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE : Node , SCREAMING_SNAKE_CASE : Node ): '''simple docstring''' UpperCamelCase__ : str = node UpperCamelCase__ : int = node.previous if node.get_previous() is None: UpperCamelCase__ : Union[str, Any] = node_to_insert else: UpperCamelCase__ : str = node_to_insert UpperCamelCase__ : str = node_to_insert def __lowercase ( self : Tuple , SCREAMING_SNAKE_CASE : Node , SCREAMING_SNAKE_CASE : Node ): '''simple docstring''' UpperCamelCase__ : Dict = node UpperCamelCase__ : Tuple = node.next if node.get_next() is None: UpperCamelCase__ : Union[str, Any] = node_to_insert else: UpperCamelCase__ : List[Any] = node_to_insert UpperCamelCase__ : Tuple = node_to_insert def __lowercase ( self : Dict , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' UpperCamelCase__ : Optional[int] = 1 UpperCamelCase__ : Any = Node(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[Any] = self.head while node: if current_position == position: self.insert_before_node(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return current_position += 1 UpperCamelCase__ : Dict = node.next self.insert_after_node(self.tail , SCREAMING_SNAKE_CASE ) def __lowercase ( self : List[Any] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' UpperCamelCase__ : List[Any] = self.head while node: if node.get_data() == item: return node UpperCamelCase__ : List[str] = node.get_next() raise Exception("Node not found" ) def __lowercase ( self : str , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if (node := self.get_node(SCREAMING_SNAKE_CASE )) is not None: if node == self.head: UpperCamelCase__ : Tuple = self.head.get_next() if node == self.tail: UpperCamelCase__ : int = self.tail.get_previous() self.remove_node_pointers(SCREAMING_SNAKE_CASE ) @staticmethod def __lowercase ( SCREAMING_SNAKE_CASE : Node ): '''simple docstring''' if node.get_next(): UpperCamelCase__ : int = node.previous if node.get_previous(): UpperCamelCase__ : Optional[Any] = node.next UpperCamelCase__ : List[Any] = None UpperCamelCase__ : Any = None def __lowercase ( self : Dict ): '''simple docstring''' return self.head is None def SCREAMING_SNAKE_CASE ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
189
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __a ( A__ ): _lowerCAmelCase : str = '''facebook/bart-large-mnli''' _lowerCAmelCase : Tuple = ( '''This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ''' '''should be the text to classify, and `labels`, which should be the list of labels to use for classification. ''' '''It returns the most likely label in the list of provided `labels` for the input text.''' ) _lowerCAmelCase : Any = '''text_classifier''' _lowerCAmelCase : int = AutoTokenizer _lowerCAmelCase : Union[str, Any] = AutoModelForSequenceClassification _lowerCAmelCase : Union[str, Any] = ['''text''', ['''text''']] _lowerCAmelCase : Dict = ['''text'''] def __lowercase ( self : int ): '''simple docstring''' super().setup() UpperCamelCase__ : Dict = self.model.config UpperCamelCase__ : Union[str, Any] = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("entail" ): UpperCamelCase__ : List[str] = int(SCREAMING_SNAKE_CASE ) if self.entailment_id == -1: raise ValueError("Could not determine the entailment ID from the model config, please pass it at init." ) def __lowercase ( self : Tuple , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' UpperCamelCase__ : Any = labels return self.pre_processor( [text] * len(SCREAMING_SNAKE_CASE ) , [F'This example is {label}' for label in labels] , return_tensors="pt" , padding="max_length" , ) def __lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' UpperCamelCase__ : List[Any] = outputs.logits UpperCamelCase__ : Any = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
189
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowercase : Tuple = ['''gpt2'''] lowercase : List[Any] = '''gpt2''' if is_tf_available(): class UpperCAmelCase_ ( tf.Module ): '''simple docstring''' def __init__( self , _SCREAMING_SNAKE_CASE ) -> Tuple: super().__init__() snake_case_ : Optional[Any] = tokenizer snake_case_ : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) snake_case_ : int = TFGPTaLMHeadModel.from_config(_SCREAMING_SNAKE_CASE ) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name="text" ),) ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: snake_case_ : Union[str, Any] = self.tokenizer(_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[int] = tokenized["input_ids"].to_tensor() snake_case_ : Optional[int] = tf.cast(input_ids_dense > 0 , tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) snake_case_ : Dict = self.model(input_ids=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE )["logits"] return outputs @require_tf @require_keras_nlp class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> Optional[Any]: super().setUp() snake_case_ : List[str] = [GPTaTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) for checkpoint in (TOKENIZER_CHECKPOINTS)] snake_case_ : List[str] = [TFGPTaTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) snake_case_ : Tuple = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] snake_case_ : List[str] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def _lowerCAmelCase ( self ) -> str: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in self.test_sentences: snake_case_ : Any = tokenizer([test_inputs] , return_tensors="tf" ) snake_case_ : str = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors snake_case_ : Any = python_outputs[key].numpy() snake_case_ : Optional[int] = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(_SCREAMING_SNAKE_CASE , tf.intaa ) == tf_outputs_values ) ) @slow def _lowerCAmelCase ( self ) -> Dict: for tf_tokenizer in self.tf_tokenizers: snake_case_ : List[str] = tf.function(_SCREAMING_SNAKE_CASE ) for test_inputs in self.test_sentences: snake_case_ : str = tf.constant(_SCREAMING_SNAKE_CASE ) snake_case_ : Tuple = compiled_tokenizer(_SCREAMING_SNAKE_CASE ) snake_case_ : List[Any] = tf_tokenizer(_SCREAMING_SNAKE_CASE ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def _lowerCAmelCase ( self ) -> Any: for tf_tokenizer in self.tf_tokenizers: snake_case_ : Optional[int] = ModelToSave(tokenizer=_SCREAMING_SNAKE_CASE ) snake_case_ : Any = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ : Optional[int] = model.serving(_SCREAMING_SNAKE_CASE ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: snake_case_ : int = Path(_SCREAMING_SNAKE_CASE ) / "saved.model" tf.saved_model.save(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , signatures={"serving_default": model.serving} ) snake_case_ : int = tf.saved_model.load(_SCREAMING_SNAKE_CASE ) snake_case_ : Tuple = loaded_model.signatures["serving_default"](_SCREAMING_SNAKE_CASE )["output_0"] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def _lowerCAmelCase ( self ) -> int: for tf_tokenizer in self.tf_tokenizers: snake_case_ : str = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ : Tuple = tf_tokenizer(_SCREAMING_SNAKE_CASE ) # Build model with some sample inputs snake_case_ : Optional[int] = tf_tokenizer.get_config() snake_case_ : List[Any] = TFGPTaTokenizer.from_config(_SCREAMING_SNAKE_CASE ) snake_case_ : Dict = model_from_config(_SCREAMING_SNAKE_CASE ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def _lowerCAmelCase ( self ) -> Optional[int]: for tf_tokenizer in self.tf_tokenizers: # for the test to run snake_case_ : int = 12_3123 for max_length in [3, 5, 1024]: snake_case_ : List[Any] = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ : List[str] = tf_tokenizer(_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) snake_case_ : List[str] = out["input_ids"].numpy().shape[1] assert out_length == max_length
360
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable lowercase : Any = {'''configuration_gpt_neox''': ['''GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXConfig''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ['''GPTNeoXTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[str] = [ '''GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXForCausalLM''', '''GPTNeoXForQuestionAnswering''', '''GPTNeoXForSequenceClassification''', '''GPTNeoXForTokenClassification''', '''GPTNeoXLayer''', '''GPTNeoXModel''', '''GPTNeoXPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
36
0
"""simple docstring""" import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets lowerCAmelCase__ = datasets.logging.get_logger(__name__) lowerCAmelCase__ = '''\ @inproceedings{bleurt, title={BLEURT: Learning Robust Metrics for Text Generation}, author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh}, booktitle={ACL}, year={2020}, url={https://arxiv.org/abs/2004.04696} } ''' lowerCAmelCase__ = '''\ BLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018) and then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune it for your specific application (the latter is expected to perform better). See the project\'s README at https://github.com/google-research/bleurt#readme for more information. ''' lowerCAmelCase__ = ''' BLEURT score. Args: `predictions` (list of str): prediction/candidate sentences `references` (list of str): reference sentences `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None. Returns: \'scores\': List of scores. Examples: >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> bleurt = datasets.load_metric("bleurt") >>> results = bleurt.compute(predictions=predictions, references=references) >>> print([round(v, 2) for v in results["scores"]]) [1.03, 1.04] ''' lowerCAmelCase__ = { '''bleurt-tiny-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip''', '''bleurt-tiny-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip''', '''bleurt-base-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip''', '''bleurt-base-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip''', '''bleurt-large-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip''', '''bleurt-large-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip''', '''BLEURT-20-D3''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip''', '''BLEURT-20-D6''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip''', '''BLEURT-20-D12''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip''', '''BLEURT-20''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip''', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): """simple docstring""" def lowercase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/google-research/bleurt" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/bleurt"] , reference_urls=["https://github.com/google-research/bleurt", "https://arxiv.org/abs/2004.04696"] , ) def lowercase__ ( self , snake_case__ ): """simple docstring""" if self.config_name == "default": logger.warning( "Using default BLEURT-Base checkpoint for sequence maximum length 128. " "You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512')." ) lowerCAmelCase : Union[str, Any] = "bleurt-base-128" if self.config_name.lower() in CHECKPOINT_URLS: lowerCAmelCase : Any = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: lowerCAmelCase : List[Any] = self.config_name.upper() else: raise KeyError( f"""{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}""" ) # download the model checkpoint specified by self.config_name and set up the scorer lowerCAmelCase : Any = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) lowerCAmelCase : int = score.BleurtScorer(os.path.join(__lowerCamelCase , __lowerCamelCase ) ) def lowercase__ ( self , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = self.scorer.score(references=__lowerCamelCase , candidates=__lowerCamelCase ) return {"scores": scores}
108
import doctest from collections import deque import numpy as np class _a : """simple docstring""" def __init__( self: Union[str, Any] ): '''simple docstring''' UpperCamelCase__: int = [2, 1, 2, -1] UpperCamelCase__: Dict = [1, 2, 3, 4] def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' UpperCamelCase__: str = len(self.first_signal ) UpperCamelCase__: Optional[Any] = len(self.second_signal ) UpperCamelCase__: str = max(__lowerCamelCase , __lowerCamelCase ) # create a zero matrix of max_length x max_length UpperCamelCase__: List[str] = [[0] * max_length for i in range(__lowerCamelCase )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(__lowerCamelCase ): UpperCamelCase__: Union[str, Any] = deque(self.second_signal ) rotated_signal.rotate(__lowerCamelCase ) for j, item in enumerate(__lowerCamelCase ): matrix[i][j] += item # multiply the matrix with the first signal UpperCamelCase__: int = np.matmul(np.transpose(__lowerCamelCase ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(__lowerCamelCase , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
149
0
'''simple docstring''' class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Tuple , UpperCamelCase__ : str = "" , UpperCamelCase__ : bool = False ): """simple docstring""" UpperCamelCase = {} # A node will be a leaf if the tree contains its word UpperCamelCase = is_leaf UpperCamelCase = prefix def A ( self : str , UpperCamelCase__ : str ): """simple docstring""" UpperCamelCase = 0 for q, w in zip(self.prefix , lowerCamelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def A ( self : Tuple , UpperCamelCase__ : list[str] ): """simple docstring""" for word in words: self.insert(lowerCamelCase__ ) def A ( self : Any , UpperCamelCase__ : str ): """simple docstring""" if self.prefix == word: UpperCamelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: UpperCamelCase = RadixNode(prefix=lowerCamelCase__ , is_leaf=lowerCamelCase__ ) else: UpperCamelCase = self.nodes[word[0]] UpperCamelCase = incoming_node.match( lowerCamelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCamelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: UpperCamelCase = remaining_prefix UpperCamelCase = self.nodes[matching_string[0]] UpperCamelCase = RadixNode(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase = aux_node if remaining_word == "": UpperCamelCase = True else: self.nodes[matching_string[0]].insert(lowerCamelCase__ ) def A ( self : int , UpperCamelCase__ : str ): """simple docstring""" UpperCamelCase = self.nodes.get(word[0] , lowerCamelCase__ ) if not incoming_node: return False else: UpperCamelCase = incoming_node.match( lowerCamelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCamelCase__ ) def A ( self : Tuple , UpperCamelCase__ : str ): """simple docstring""" UpperCamelCase = self.nodes.get(word[0] , lowerCamelCase__ ) if not incoming_node: return False else: UpperCamelCase = incoming_node.match( lowerCamelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCamelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: UpperCamelCase = list(self.nodes.values() )[0] UpperCamelCase = merging_node.is_leaf self.prefix += merging_node.prefix UpperCamelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: UpperCamelCase = False # If there is 1 edge, we merge it with its child else: UpperCamelCase = list(incoming_node.nodes.values() )[0] UpperCamelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix UpperCamelCase = merging_node.nodes return True def A ( self : List[str] , UpperCamelCase__ : int = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __lowerCamelCase ( ) -> List[Any]: """simple docstring""" UpperCamelCase = '''banana bananas bandana band apple all beast'''.split() UpperCamelCase = RadixNode() root.insert_many(A__ ) assert all(root.find(A__ ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __lowerCamelCase ( ) -> int: """simple docstring""" assert test_trie() def __lowerCamelCase ( ) -> Dict: """simple docstring""" UpperCamelCase = RadixNode() UpperCamelCase = '''banana bananas bandanas bandana band apple all beast'''.split() root.insert_many(A__ ) print('Words:' , A__ ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
357
'''simple docstring''' def __lowerCamelCase ( A__ , A__ ) -> float: """simple docstring""" if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modulus' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
249
0
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): UpperCAmelCase__ = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right UpperCAmelCase__ = 1_2_8_0_2_2 UpperCAmelCase__ = 1_2_8_0_2_8 @require_sentencepiece class a ( lowerCAmelCase_ , unittest.TestCase ): _snake_case : List[Any] = MaMaaaTokenizer _snake_case : int = False _snake_case : List[str] = False _snake_case : Any = True def lowerCAmelCase_ ( self : List[Any] ): super().setUp() _UpperCAmelCase = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] _UpperCAmelCase = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) _UpperCAmelCase = Path(self.tmpdirname ) save_json(__lowerCAmelCase , save_dir / VOCAB_FILES_NAMES["""vocab_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(__lowerCAmelCase , save_dir / VOCAB_FILES_NAMES["""spm_file"""] ) _UpperCAmelCase = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase_ ( self : Union[str, Any] , **__lowerCAmelCase : str ): return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase ) def lowerCAmelCase_ ( self : Dict , __lowerCAmelCase : Any ): return ( "This is a test", "This is a test", ) def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = """</s>""" _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCAmelCase ) , __lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCAmelCase ) , __lowerCAmelCase ) def lowerCAmelCase_ ( self : int ): _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<s>""" ) self.assertEqual(len(__lowerCAmelCase ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("""Skip this test while all models are still to be uploaded.""" ) def lowerCAmelCase_ ( self : Tuple ): pass def lowerCAmelCase_ ( self : str ): _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__lowerCAmelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , [2, 3, 4, 5, 6] , ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(__lowerCAmelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) _UpperCAmelCase = tokenizer.convert_tokens_to_string(__lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , """This is a test""" ) @slow def lowerCAmelCase_ ( self : Optional[Any] ): # fmt: off _UpperCAmelCase = {"""input_ids""": [[12_8022, 11_0108, 397, 11, 3_8272, 2247, 12_4811, 285, 1_8105, 1586, 207, 7, 3_9534, 4428, 397, 1019, 1_8105, 1586, 207, 7, 4_1337, 1_6786, 241, 7, 2_0214, 17, 12_5690, 1_0398, 7, 4_4378, 5_8069, 6_8342, 7798, 7343, 11, 299, 3_3310, 4, 158, 3_7350, 9_4077, 4569, 299, 3_3310, 90, 4, 5_2840, 290, 4, 3_1270, 112, 299, 682, 4, 5_2840, 3_9953, 1_4079, 193, 5_2519, 9_0894, 1_7894, 12_0697, 11, 4_0445, 551, 17, 1019, 5_2519, 9_0894, 1_7756, 963, 11, 4_0445, 480, 17, 9792, 1120, 5173, 1393, 6240, 1_6786, 241, 12_0996, 28, 1245, 1393, 11_8240, 1_1123, 1019, 9_3612, 2691, 1_0618, 9_8058, 12_0409, 1928, 279, 4, 4_0683, 367, 178, 207, 1019, 103, 10_3121, 506, 6_5296, 5, 2], [12_8022, 2_1217, 367, 117, 12_5450, 128, 719, 7, 7308, 40, 9_3612, 1_2669, 1116, 1_6704, 71, 1_7785, 3699, 1_5592, 35, 144, 9584, 241, 1_1943, 713, 950, 799, 2247, 8_8427, 150, 149, 11_8813, 12_0706, 1019, 10_6906, 8_1518, 28, 1224, 2_2799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [12_8022, 1658, 12_3311, 5155, 5578, 4722, 279, 1_4947, 2366, 1120, 1197, 14, 1348, 9232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCAmelCase , model_name="""facebook/m2m100_418M""" , revision="""c168bae485c864188cf9aa0e4108b0b6934dc91e""" , ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): _snake_case : List[str] = 'facebook/m2m100_418M' _snake_case : Optional[int] = [ 'In my opinion, there are two levels of response from the French government.', 'NSA Affair Emphasizes Complete Lack of Debate on Intelligence', ] _snake_case : int = [ 'Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.', 'L\'affaire NSA souligne l\'absence totale de débat sur le renseignement', ] # fmt: off _snake_case : Optional[int] = [EN_CODE, 5_93, 19_49, 11_57_81, 4, 7_15_86, 42_34, 6_06_33, 12_62_33, 4_32, 12_38_08, 1_55_92, 11_97, 11_71_32, 12_06_18, 5, 2] @classmethod def lowerCAmelCase_ ( cls : Any ): _UpperCAmelCase = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en""" , tgt_lang="""fr""" ) _UpperCAmelCase = 1 return cls def lowerCAmelCase_ ( self : Dict ): self.assertEqual(self.tokenizer.get_lang_id("""ar""" ) , 12_8006 ) self.assertEqual(self.tokenizer.get_lang_id("""en""" ) , 12_8022 ) self.assertEqual(self.tokenizer.get_lang_id("""ro""" ) , 12_8076 ) self.assertEqual(self.tokenizer.get_lang_id("""mr""" ) , 12_8063 ) def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase = self.tokenizer.get_vocab() self.assertEqual(len(__lowerCAmelCase ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["""<unk>"""] , 3 ) self.assertIn(self.tokenizer.get_lang_token("""en""" ) , __lowerCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): _UpperCAmelCase = """en""" _UpperCAmelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowerCAmelCase ) def lowerCAmelCase_ ( self : int ): self.assertIn(__lowerCAmelCase , self.tokenizer.all_special_ids ) # fmt: off _UpperCAmelCase = [FR_CODE, 5364, 82, 8642, 4, 294, 47, 8, 1_4028, 136, 3286, 9706, 6, 9_0797, 6, 14_4012, 162, 8_8128, 3_0061, 5, 2] # fmt: on _UpperCAmelCase = self.tokenizer.decode(__lowerCAmelCase , skip_special_tokens=__lowerCAmelCase ) _UpperCAmelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) self.assertNotIn(self.tokenizer.eos_token , __lowerCAmelCase ) def lowerCAmelCase_ ( self : Any ): _UpperCAmelCase = tempfile.mkdtemp() _UpperCAmelCase = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(__lowerCAmelCase ) _UpperCAmelCase = MaMaaaTokenizer.from_pretrained(__lowerCAmelCase ) self.assertDictEqual(new_tok.lang_token_to_id , __lowerCAmelCase ) @require_torch def lowerCAmelCase_ ( self : str ): _UpperCAmelCase = """en""" _UpperCAmelCase = """fr""" _UpperCAmelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowerCAmelCase , return_tensors="""pt""" ) _UpperCAmelCase = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _UpperCAmelCase = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = """mr""" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _UpperCAmelCase = """zh""" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def lowerCAmelCase_ ( self : int ): _UpperCAmelCase = """mr""" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _UpperCAmelCase = """zh""" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def lowerCAmelCase_ ( self : Any ): _UpperCAmelCase = self.tokenizer._build_translation_inputs("""A test""" , return_tensors="""pt""" , src_lang="""en""" , tgt_lang="""ar""" ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , { # en_XX, A, test, EOS """input_ids""": [[12_8022, 58, 4183, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 12_8006, } , )
289
"""simple docstring""" import warnings warnings.warn( """memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: """ """`from accelerate import find_executable_batch_size` to avoid this warning.""", FutureWarning, )
289
1
import argparse import copy def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[Any]: __lowerCamelCase : Optional[Any] = {} with open(a__ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __lowerCamelCase : int = [] _list.append([line.split()[1], line.split()[2]] ) __lowerCamelCase : Any = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __lowerCamelCase : int = [] _list.append([line.split()[0], line.split()[2]] ) __lowerCamelCase : Any = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: with open(a__ ) as f: __lowerCamelCase : Tuple = f.read(1 ) __lowerCamelCase : Any = start_node __lowerCamelCase : List[Any] = [] __lowerCamelCase : str = start_node __lowerCamelCase : int = 0 while visiting not in first_solution: __lowerCamelCase : Dict = 1_0_0_0_0 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a__ ) and k[0] not in first_solution: __lowerCamelCase : List[str] = k[1] __lowerCamelCase : int = k[0] first_solution.append(a__ ) __lowerCamelCase : Optional[Any] = distance_of_first_solution + int(a__ ) __lowerCamelCase : Optional[Any] = best_node first_solution.append(a__ ) __lowerCamelCase : Optional[Any] = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __lowerCamelCase : Union[str, Any] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0_0_0_0 ) return first_solution, distance_of_first_solution def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Union[str, Any]: __lowerCamelCase : List[Any] = [] for n in solution[1:-1]: __lowerCamelCase : List[str] = solution.index(a__ ) for kn in solution[1:-1]: __lowerCamelCase : int = solution.index(a__ ) if n == kn: continue __lowerCamelCase : Tuple = copy.deepcopy(a__ ) __lowerCamelCase : Tuple = kn __lowerCamelCase : Dict = n __lowerCamelCase : str = 0 for k in _tmp[:-1]: __lowerCamelCase : List[Any] = _tmp[_tmp.index(a__ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __lowerCamelCase : Any = distance + int(i[1] ) _tmp.append(a__ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __lowerCamelCase : Union[str, Any] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowerCamelCase__ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Any: __lowerCamelCase : str = 1 __lowerCamelCase : List[str] = first_solution __lowerCamelCase : Tuple = [] __lowerCamelCase : Optional[int] = distance_of_first_solution __lowerCamelCase : Dict = solution while count <= iters: __lowerCamelCase : Union[str, Any] = find_neighborhood(a__ , a__ ) __lowerCamelCase : Optional[Any] = 0 __lowerCamelCase : Optional[Any] = neighborhood[index_of_best_solution] __lowerCamelCase : Union[str, Any] = len(a__ ) - 1 __lowerCamelCase : List[str] = False while not found: __lowerCamelCase : Optional[Any] = 0 while i < len(a__ ): if best_solution[i] != solution[i]: __lowerCamelCase : Optional[int] = best_solution[i] __lowerCamelCase : int = solution[i] break __lowerCamelCase : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __lowerCamelCase : Tuple = True __lowerCamelCase : Optional[int] = best_solution[:-1] __lowerCamelCase : str = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __lowerCamelCase : Optional[Any] = cost __lowerCamelCase : int = solution else: __lowerCamelCase : Dict = index_of_best_solution + 1 __lowerCamelCase : Tuple = neighborhood[index_of_best_solution] if len(a__ ) >= size: tabu_list.pop(0 ) __lowerCamelCase : int = count + 1 return best_solution_ever, best_cost def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__=None ) -> str: __lowerCamelCase : str = generate_neighbours(args.File ) __lowerCamelCase , __lowerCamelCase : List[Any] = generate_first_solution( args.File , a__ ) __lowerCamelCase , __lowerCamelCase : Tuple = tabu_search( a__ , a__ , a__ , args.Iterations , args.Size , ) print(F"Best solution: {best_sol}, with total distance: {best_cost}." ) if __name__ == "__main__": a =argparse.ArgumentParser(description="""Tabu Search""") parser.add_argument( """-f""", """--File""", type=str, help="""Path to the file containing the data""", required=True, ) parser.add_argument( """-i""", """--Iterations""", type=int, help="""How many iterations the algorithm should perform""", required=True, ) parser.add_argument( """-s""", """--Size""", type=int, help="""Size of the tabu list""", required=True ) # Pass the arguments to main method main(parser.parse_args())
361
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py a ="""src/transformers""" a ="""docs/source/en""" a =""".""" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Optional[int]: with open(lowerCamelCase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: __lowerCamelCase : Any = f.readlines() # Find the start prompt. __lowerCamelCase : List[str] = 0 while not lines[start_index].startswith(lowerCamelCase__ ): start_index += 1 start_index += 1 __lowerCamelCase : int = start_index while not lines[end_index].startswith(lowerCamelCase__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | a ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. a =re.compile(r"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") a =re.compile(r"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. a =re.compile(r"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. a =direct_transformers_import(TRANSFORMERS_PATH) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[str]: __lowerCamelCase : int = re.finditer('.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)' , lowerCamelCase__ ) return [m.group(0 ) for m in matches] def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Tuple: __lowerCamelCase : int = 2 if text == '✅' or text == '❌' else len(lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = (width - text_length) // 2 __lowerCamelCase : List[Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def SCREAMING_SNAKE_CASE__ ( ) -> str: __lowerCamelCase : Union[str, Any] = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES __lowerCamelCase : List[str] = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } __lowerCamelCase : Dict = {name: config.replace('Config' , '' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. __lowerCamelCase : Union[str, Any] = collections.defaultdict(lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = collections.defaultdict(lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = collections.defaultdict(lowerCamelCase__ ) __lowerCamelCase : List[str] = collections.defaultdict(lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = collections.defaultdict(lowerCamelCase__ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCamelCase__ ): __lowerCamelCase : List[Any] = None if attr_name.endswith('Tokenizer' ): __lowerCamelCase : Dict = slow_tokenizers __lowerCamelCase : List[Any] = attr_name[:-9] elif attr_name.endswith('TokenizerFast' ): __lowerCamelCase : Union[str, Any] = fast_tokenizers __lowerCamelCase : str = attr_name[:-1_3] elif _re_tf_models.match(lowerCamelCase__ ) is not None: __lowerCamelCase : List[str] = tf_models __lowerCamelCase : Optional[int] = _re_tf_models.match(lowerCamelCase__ ).groups()[0] elif _re_flax_models.match(lowerCamelCase__ ) is not None: __lowerCamelCase : List[Any] = flax_models __lowerCamelCase : Optional[Any] = _re_flax_models.match(lowerCamelCase__ ).groups()[0] elif _re_pt_models.match(lowerCamelCase__ ) is not None: __lowerCamelCase : Optional[int] = pt_models __lowerCamelCase : Any = _re_pt_models.match(lowerCamelCase__ ).groups()[0] if lookup_dict is not None: while len(lowerCamelCase__ ) > 0: if attr_name in model_name_to_prefix.values(): __lowerCamelCase : List[Any] = True break # Try again after removing the last word in the name __lowerCamelCase : str = ''.join(camel_case_split(lowerCamelCase__ )[:-1] ) # Let's build that table! __lowerCamelCase : str = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) __lowerCamelCase : Union[str, Any] = ['Model', 'Tokenizer slow', 'Tokenizer fast', 'PyTorch support', 'TensorFlow support', 'Flax Support'] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). __lowerCamelCase : List[Any] = [len(lowerCamelCase__ ) + 2 for c in columns] __lowerCamelCase : int = max([len(lowerCamelCase__ ) for name in model_names] ) + 2 # Build the table per se __lowerCamelCase : Union[str, Any] = '|' + '|'.join([_center_text(lowerCamelCase__ , lowerCamelCase__ ) for c, w in zip(lowerCamelCase__ , lowerCamelCase__ )] ) + '|\n' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([':' + '-' * (w - 2) + ':' for w in widths] ) + "|\n" __lowerCamelCase : List[str] = {True: '✅', False: '❌'} for name in model_names: __lowerCamelCase : Optional[int] = model_name_to_prefix[name] __lowerCamelCase : Any = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCamelCase__ , lowerCamelCase__ ) for l, w in zip(lowerCamelCase__ , lowerCamelCase__ )] ) + "|\n" return table def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__=False ) -> Any: __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : List[str] = _find_text_in_file( filename=os.path.join(lowerCamelCase__ , 'index.md' ) , start_prompt='<!--This table is updated automatically from the auto modules' , end_prompt='<!-- End table-->' , ) __lowerCamelCase : List[Any] = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCamelCase__ , 'index.md' ) , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( 'The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.' ) if __name__ == "__main__": a =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") a =parser.parse_args() check_model_table(args.fix_and_overwrite)
113
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ : int = { '''configuration_lilt''': ['''LILT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LiltConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : int = [ '''LILT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LiltForQuestionAnswering''', '''LiltForSequenceClassification''', '''LiltForTokenClassification''', '''LiltModel''', '''LiltPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys lowerCAmelCase__ : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
143
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Tuple = logging.get_logger(__name__) lowerCAmelCase__ : Union[str, Any] = {'''vocab_file''': '''sentencepiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ : Dict = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, '''tokenizer_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ : List[Any] = { '''google/rembert''': 2_56, } lowerCAmelCase__ : List[str] = '''▁''' class __snake_case ( _lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RemBertTokenizer def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase="[CLS]" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<unk>" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<pad>" , __UpperCamelCase="[CLS]" , __UpperCamelCase="[MASK]" , **__UpperCamelCase , ) -> Tuple: '''simple docstring''' snake_case__ : Optional[Any] = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token super().__init__( __UpperCamelCase , tokenizer_file=__UpperCamelCase , do_lower_case=__UpperCamelCase , remove_space=__UpperCamelCase , keep_accents=__UpperCamelCase , bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , pad_token=__UpperCamelCase , cls_token=__UpperCamelCase , mask_token=__UpperCamelCase , **__UpperCamelCase , ) snake_case__ : int = do_lower_case snake_case__ : Any = remove_space snake_case__ : List[Any] = keep_accents snake_case__ : Dict = vocab_file snake_case__ : int = False if not self.vocab_file else True def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : Dict = [self.sep_token_id] snake_case__ : List[str] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__UpperCamelCase )) + [1] + ([0] * len(__UpperCamelCase )) + [1] return [1] + ([0] * len(__UpperCamelCase )) + [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : List[Any] = [self.sep_token_id] snake_case__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(__UpperCamelCase ): logger.error('Vocabulary path ({}) should be a directory'.format(__UpperCamelCase ) ) return snake_case__ : List[str] = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ): copyfile(self.vocab_file , __UpperCamelCase ) return (out_vocab_file,)
143
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __SCREAMING_SNAKE_CASE ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :int = None SCREAMING_SNAKE_CASE__ :int = BloomTokenizerFast SCREAMING_SNAKE_CASE__ :Tuple = BloomTokenizerFast SCREAMING_SNAKE_CASE__ :Dict = True SCREAMING_SNAKE_CASE__ :Dict = False SCREAMING_SNAKE_CASE__ :Optional[int] = "tokenizer_file" SCREAMING_SNAKE_CASE__ :int = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: super().setUp() _UpperCamelCase : int = BloomTokenizerFast.from_pretrained("bigscience/tokenizer" ) tokenizer.save_pretrained(self.tmpdirname ) def __SCREAMING_SNAKE_CASE ( self : str , **__a : Tuple ) -> List[str]: kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **__a ) def __SCREAMING_SNAKE_CASE ( self : str ) -> Any: _UpperCamelCase : int = self.get_rust_tokenizer() _UpperCamelCase : List[Any] = ["The quick brown fox</s>", "jumps over the lazy dog</s>"] _UpperCamelCase : int = [[2175, 2_3714, 7_3173, 14_4252, 2], [77, 13_2619, 3478, 368, 10_9586, 3_5433, 2]] _UpperCamelCase : List[str] = tokenizer.batch_encode_plus(__a )["input_ids"] self.assertListEqual(__a , __a ) _UpperCamelCase : List[Any] = tokenizer.batch_decode(__a ) self.assertListEqual(__a , __a ) def __SCREAMING_SNAKE_CASE ( self : Dict , __a : Optional[int]=6 ) -> Dict: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): _UpperCamelCase : str = self.rust_tokenizer_class.from_pretrained(__a , **__a ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input _UpperCamelCase : str = "This is a simple input" _UpperCamelCase : Union[str, Any] = ["This is a simple input 1", "This is a simple input 2"] _UpperCamelCase : Tuple = ("This is a simple input", "This is a pair") _UpperCamelCase : List[Any] = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests try: tokenizer_r.encode(__a , max_length=__a ) tokenizer_r.encode_plus(__a , max_length=__a ) tokenizer_r.batch_encode_plus(__a , max_length=__a ) tokenizer_r.encode(__a , max_length=__a ) tokenizer_r.batch_encode_plus(__a , max_length=__a ) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding" ) _UpperCamelCase : Tuple = None # Hotfixing padding = None self.assertRaises(__a , tokenizer_r.encode , __a , max_length=__a , padding="max_length" ) # Simple input self.assertRaises(__a , tokenizer_r.encode_plus , __a , max_length=__a , padding="max_length" ) # Simple input self.assertRaises( __a , tokenizer_r.batch_encode_plus , __a , max_length=__a , padding="max_length" , ) # Pair input self.assertRaises(__a , tokenizer_r.encode , __a , max_length=__a , padding="max_length" ) # Pair input self.assertRaises(__a , tokenizer_r.encode_plus , __a , max_length=__a , padding="max_length" ) # Pair input self.assertRaises( __a , tokenizer_r.batch_encode_plus , __a , max_length=__a , padding="max_length" , ) def __SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: _UpperCamelCase : Dict = self.get_rust_tokenizer() _UpperCamelCase : Optional[Any] = load_dataset("xnli" , "all_languages" , split="test" , streaming=__a ) _UpperCamelCase : List[str] = next(iter(__a ) )["premise"] # pick up one data _UpperCamelCase : List[str] = list(sample_data.values() ) _UpperCamelCase : Any = list(map(tokenizer.encode , __a ) ) _UpperCamelCase : List[Any] = [tokenizer.decode(__a , clean_up_tokenization_spaces=__a ) for x in output_tokens] self.assertListEqual(__a , __a ) def __SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
310
"""simple docstring""" from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent lowerCamelCase__ = {"UserAgent": UserAgent().random} def lowercase__ ( lowercase_ ) -> dict: """simple docstring""" _UpperCamelCase : str = script.contents[0] _UpperCamelCase : Any = json.loads(data[data.find("{\"config\"" ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Dict , __a : str ) -> Tuple: _UpperCamelCase : List[str] = F'''https://www.instagram.com/{username}/''' _UpperCamelCase : Optional[Any] = self.get_json() def __SCREAMING_SNAKE_CASE ( self : Tuple ) -> dict: _UpperCamelCase : int = requests.get(self.url , headers=__a ).text _UpperCamelCase : Union[str, Any] = BeautifulSoup(__a , "html.parser" ).find_all("script" ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : List[Any] ) -> str: return F'''{self.__class__.__name__}(\'{self.username}\')''' def __str__( self : str ) -> str: return F'''{self.fullname} ({self.username}) is {self.biography}''' @property def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: return self.user_data["username"] @property def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> str: return self.user_data["full_name"] @property def __SCREAMING_SNAKE_CASE ( self : Dict ) -> str: return self.user_data["biography"] @property def __SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> str: return self.user_data["business_email"] @property def __SCREAMING_SNAKE_CASE ( self : Any ) -> str: return self.user_data["external_url"] @property def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: return self.user_data["edge_followed_by"]["count"] @property def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: return self.user_data["edge_follow"]["count"] @property def __SCREAMING_SNAKE_CASE ( self : Dict ) -> int: return self.user_data["edge_owner_to_timeline_media"]["count"] @property def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> str: return self.user_data["profile_pic_url_hd"] @property def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> bool: return self.user_data["is_verified"] @property def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> bool: return self.user_data["is_private"] def lowercase__ ( lowercase_ = "github" ) -> None: """simple docstring""" import os if os.environ.get("CI" ): return # test failing on GitHub Actions _UpperCamelCase : Union[str, Any] = InstagramUser(lowercase_ ) assert instagram_user.user_data assert isinstance(instagram_user.user_data ,lowercase_ ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 120_000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith("https://instagram." ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = InstagramUser("github") print(instagram_user) print(f"""{instagram_user.number_of_posts = }""") print(f"""{instagram_user.number_of_followers = }""") print(f"""{instagram_user.number_of_followings = }""") print(f"""{instagram_user.email = }""") print(f"""{instagram_user.website = }""") print(f"""{instagram_user.profile_picture_url = }""") print(f"""{instagram_user.is_verified = }""") print(f"""{instagram_user.is_private = }""")
310
1
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( A__ ): A__ = ['image_processor', 'tokenizer'] A__ = 'BlipImageProcessor' A__ = ('BertTokenizer', 'BertTokenizerFast') def __init__( self : Optional[int] , _a : Dict , _a : List[Any] ) -> Any: '''simple docstring''' _SCREAMING_SNAKE_CASE =False super().__init__(__UpperCAmelCase , __UpperCAmelCase ) _SCREAMING_SNAKE_CASE =self.image_processor def __call__( self : Union[str, Any] , _a : ImageInput = None , _a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , _a : bool = True , _a : Union[bool, str, PaddingStrategy] = False , _a : Union[bool, str, TruncationStrategy] = None , _a : Optional[int] = None , _a : int = 0 , _a : Optional[int] = None , _a : Optional[bool] = None , _a : bool = False , _a : bool = False , _a : bool = False , _a : bool = False , _a : bool = False , _a : bool = True , _a : Optional[Union[str, TensorType]] = None , **_a : Any , ) -> BatchEncoding: '''simple docstring''' if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: _SCREAMING_SNAKE_CASE =self.tokenizer _SCREAMING_SNAKE_CASE =self.tokenizer( text=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=__UpperCAmelCase , max_length=__UpperCAmelCase , stride=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , return_overflowing_tokens=__UpperCAmelCase , return_special_tokens_mask=__UpperCAmelCase , return_offsets_mapping=__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase , return_length=__UpperCAmelCase , verbose=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) return text_encoding # add pixel_values _SCREAMING_SNAKE_CASE =self.image_processor(__UpperCAmelCase , return_tensors=__UpperCAmelCase ) if text is not None: _SCREAMING_SNAKE_CASE =self.tokenizer( text=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=__UpperCAmelCase , max_length=__UpperCAmelCase , stride=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , return_overflowing_tokens=__UpperCAmelCase , return_special_tokens_mask=__UpperCAmelCase , return_offsets_mapping=__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase , return_length=__UpperCAmelCase , verbose=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) else: _SCREAMING_SNAKE_CASE =None if text_encoding is not None: encoding_image_processor.update(__UpperCAmelCase ) return encoding_image_processor def A ( self : Any , *_a : int , **_a : Optional[int] ) -> int: '''simple docstring''' return self.tokenizer.batch_decode(*__UpperCAmelCase , **__UpperCAmelCase ) def A ( self : Union[str, Any] , *_a : Dict , **_a : Union[str, Any] ) -> List[Any]: '''simple docstring''' return self.tokenizer.decode(*__UpperCAmelCase , **__UpperCAmelCase ) @property def A ( self : Any ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.tokenizer.model_input_names _SCREAMING_SNAKE_CASE =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
47
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A_ : List[Any] = logging.get_logger(__name__) A_ : List[Any] = "▁" A_ : str = {"vocab_file": "sentencepiece.bpe.model"} A_ : Union[str, Any] = { "vocab_file": { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model" ), } } A_ : List[str] = { "xlm-roberta-base": 512, "xlm-roberta-large": 512, "xlm-roberta-large-finetuned-conll02-dutch": 512, "xlm-roberta-large-finetuned-conll02-spanish": 512, "xlm-roberta-large-finetuned-conll03-english": 512, "xlm-roberta-large-finetuned-conll03-german": 512, } class lowerCamelCase (A__ ): lowerCamelCase__ : Optional[int] = VOCAB_FILES_NAMES lowerCamelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple="<s>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="</s>" , __UpperCAmelCase : Tuple="<s>" , __UpperCAmelCase : Union[str, Any]="<unk>" , __UpperCAmelCase : List[str]="<pad>" , __UpperCAmelCase : Dict="<mask>" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Optional[int] , ) -> None: # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE__ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token SCREAMING_SNAKE_CASE__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) SCREAMING_SNAKE_CASE__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCAmelCase ) ) SCREAMING_SNAKE_CASE__ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token SCREAMING_SNAKE_CASE__ = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = len(self.sp_model ) + self.fairseq_offset SCREAMING_SNAKE_CASE__ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = self.__dict__.copy() SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , __UpperCAmelCase : Optional[int] ) -> List[str]: SCREAMING_SNAKE_CASE__ = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def SCREAMING_SNAKE_CASE ( self : Dict , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] SCREAMING_SNAKE_CASE__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : Dict , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] + ([0] * len(__UpperCAmelCase )) + [1] def SCREAMING_SNAKE_CASE ( self : int , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: SCREAMING_SNAKE_CASE__ = [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[int]: return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Tuple: SCREAMING_SNAKE_CASE__ = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE ( self : List[Any] , __UpperCAmelCase : str ) -> List[str]: return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __UpperCAmelCase : List[str] ) -> Tuple: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] SCREAMING_SNAKE_CASE__ = self.sp_model.PieceToId(__UpperCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def SCREAMING_SNAKE_CASE ( self : List[Any] , __UpperCAmelCase : int ) -> Any: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __UpperCAmelCase : List[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = """""".join(__UpperCAmelCase ).replace(__UpperCAmelCase , """ """ ).strip() return out_string def SCREAMING_SNAKE_CASE ( self : int , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return SCREAMING_SNAKE_CASE__ = os.path.join( __UpperCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , """wb""" ) as fi: SCREAMING_SNAKE_CASE__ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
165
0
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING A__ : Optional[int] = logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE_ ) class snake_case__ ( SCREAMING_SNAKE_CASE_ ): def __init__( self : Dict , **__a : Dict ) -> List[str]: '''simple docstring''' super().__init__(**__a ) if self.framework == "tf": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) requires_backends(self , 'vision' ) self.check_model_type(__a ) def __call__( self : Dict , __a : Union[str, "Image.Image", List[Dict[str, Any]]] , __a : Union[str, List[str]] = None , **__a : Any , ) -> str: '''simple docstring''' if "text_queries" in kwargs: __snake_case : Tuple = kwargs.pop('text_queries' ) if isinstance(__a , (str, Image.Image) ): __snake_case : List[str] = {'image': image, 'candidate_labels': candidate_labels} else: __snake_case : Optional[int] = image __snake_case : Union[str, Any] = super().__call__(__a , **__a ) return results def A_ ( self : List[str] , **__a : int ) -> Optional[int]: '''simple docstring''' __snake_case : str = {} if "threshold" in kwargs: __snake_case : int = kwargs['threshold'] if "top_k" in kwargs: __snake_case : int = kwargs['top_k'] return {}, {}, postprocess_params def A_ ( self : List[str] , __a : Union[str, Any] ) -> Tuple: '''simple docstring''' __snake_case : Tuple = load_image(inputs['image'] ) __snake_case : Optional[int] = inputs['candidate_labels'] if isinstance(__a , __a ): __snake_case : Dict = candidate_labels.split(',' ) __snake_case : Tuple = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(__a ): __snake_case : str = self.tokenizer(__a , return_tensors=self.framework ) __snake_case : Tuple = self.image_processor(__a , return_tensors=self.framework ) yield { "is_last": i == len(__a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def A_ ( self : Optional[Any] , __a : Optional[int] ) -> Optional[int]: '''simple docstring''' __snake_case : Any = model_inputs.pop('target_size' ) __snake_case : Tuple = model_inputs.pop('candidate_label' ) __snake_case : List[Any] = model_inputs.pop('is_last' ) __snake_case : Any = self.model(**__a ) __snake_case : Optional[int] = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def A_ ( self : Union[str, Any] , __a : Optional[Any] , __a : List[Any]=0.1 , __a : Dict=None ) -> Optional[Any]: '''simple docstring''' __snake_case : str = [] for model_output in model_outputs: __snake_case : Optional[Any] = model_output['candidate_label'] __snake_case : List[str] = BaseModelOutput(__a ) __snake_case : Optional[int] = self.image_processor.post_process_object_detection( outputs=__a , threshold=__a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): __snake_case : Dict = outputs['scores'][index].item() __snake_case : List[Any] = self._get_bounding_box(outputs['boxes'][index][0] ) __snake_case : Any = {'score': score, 'label': label, 'box': box} results.append(__a ) __snake_case : str = sorted(__a , key=lambda __a : x["score"] , reverse=__a ) if top_k: __snake_case : Any = results[:top_k] return results def A_ ( self : Optional[int] , __a : "torch.Tensor" ) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) __snake_case , __snake_case , __snake_case , __snake_case : Dict = box.int().tolist() __snake_case : str = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A__ : Tuple = logging.get_logger(__name__) A__ : Optional[int] = {} class snake_case__ ( SCREAMING_SNAKE_CASE_ ): A__ = '''llama''' A__ = ['''past_key_values'''] def __init__( self : Any , __a : List[str]=32000 , __a : Union[str, Any]=4096 , __a : Optional[Any]=11008 , __a : Any=32 , __a : str=32 , __a : Optional[int]=None , __a : Dict="silu" , __a : Dict=2048 , __a : List[str]=0.0_2 , __a : Union[str, Any]=1e-6 , __a : Dict=True , __a : List[str]=0 , __a : Tuple=1 , __a : Tuple=2 , __a : Optional[Any]=1 , __a : Any=False , __a : Tuple=None , **__a : List[Any] , ) -> Optional[int]: '''simple docstring''' __snake_case : str = vocab_size __snake_case : List[str] = max_position_embeddings __snake_case : List[Any] = hidden_size __snake_case : Union[str, Any] = intermediate_size __snake_case : Optional[int] = num_hidden_layers __snake_case : List[Any] = num_attention_heads # for backward compatibility if num_key_value_heads is None: __snake_case : Optional[int] = num_attention_heads __snake_case : Optional[Any] = num_key_value_heads __snake_case : int = hidden_act __snake_case : Any = initializer_range __snake_case : Any = rms_norm_eps __snake_case : Union[str, Any] = pretraining_tp __snake_case : Optional[int] = use_cache __snake_case : Any = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , tie_word_embeddings=__a , **__a , ) def A_ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __a ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' f'''got {self.rope_scaling}''' ) __snake_case : Optional[Any] = self.rope_scaling.get('type' , __a ) __snake_case : Tuple = self.rope_scaling.get('factor' , __a ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(__a , __a ) or rope_scaling_factor <= 1.0: raise ValueError(f'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
0
1
# Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib UpperCamelCase__ = get_logger() UpperCamelCase__ = None class a__ ( TensorFormatter[Mapping, """jax.Array""", Mapping] ): def __init__( self , _A=None , _A=None , **_A ): """simple docstring""" super().__init__(features=_A ) import jax from jaxlib.xla_client import Device if isinstance(_A , _A ): raise ValueError( f"""Expected {device} to be a `str` not {type(_A )}, as `jaxlib.xla_extension.Device` """ "is not serializable neither with `pickle` nor with `dill`. Instead you can surround " "the device with `str()` to get its string identifier that will be internally mapped " "to the actual `jaxlib.xla_extension.Device`." ) __lowerCAmelCase = device if isinstance(_A , _A ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: __lowerCAmelCase = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( f"""Device with string identifier {self.device} not listed among the available """ f"""devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default """ f"""device: {str(jax.devices()[0] )}.""" ) __lowerCAmelCase = str(jax.devices()[0] ) __lowerCAmelCase = jnp_array_kwargs @staticmethod def __SCREAMING_SNAKE_CASE( ): """simple docstring""" import jax return {str(_A ): device for device in jax.devices()} def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" import jax import jax.numpy as jnp if isinstance(_A , _A ) and column: if all( isinstance(_A , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(_A , axis=0 ) return column def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" import jax import jax.numpy as jnp if isinstance(_A , (str, bytes, type(_A )) ): return value elif isinstance(_A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() __lowerCAmelCase = {} if isinstance(_A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: __lowerCAmelCase = {"dtype": jnp.intaa} else: __lowerCAmelCase = {"dtype": jnp.intaa} elif isinstance(_A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): __lowerCAmelCase = {"dtype": jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(_A , PIL.Image.Image ): __lowerCAmelCase = np.asarray(_A ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: __lowerCAmelCase = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(_A , **{**default_dtype, **self.jnp_array_kwargs} ) def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(_A , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(_A , "__array__" ) and not isinstance(_A , jax.Array ): __lowerCAmelCase = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(_A , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(_A ) for substruct in data_struct] ) elif isinstance(_A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(_A ) for substruct in data_struct] ) return self._tensorize(_A ) def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" return map_nested(self._recursive_tensorize , _A , map_list=_A ) def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" __lowerCAmelCase = self.numpy_arrow_extractor().extract_row(_A ) __lowerCAmelCase = self.python_features_decoder.decode_row(_A ) return self.recursive_tensorize(_A ) def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" __lowerCAmelCase = self.numpy_arrow_extractor().extract_column(_A ) __lowerCAmelCase = self.python_features_decoder.decode_column(_A , pa_table.column_names[0] ) __lowerCAmelCase = self.recursive_tensorize(_A ) __lowerCAmelCase = self._consolidate(_A ) return column def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" __lowerCAmelCase = self.numpy_arrow_extractor().extract_batch(_A ) __lowerCAmelCase = self.python_features_decoder.decode_batch(_A ) __lowerCAmelCase = self.recursive_tensorize(_A ) for column_name in batch: __lowerCAmelCase = self._consolidate(batch[column_name] ) return batch
92
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __snake_case : __lowerCamelCase : str = BlenderbotConfig __lowerCamelCase : Optional[Any] = {} __lowerCamelCase : Optional[int] = """gelu""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=False , snake_case__=99 , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=20 , snake_case__=2 , snake_case__=1 , snake_case__=0 , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =parent UpperCAmelCase : Optional[int] =batch_size UpperCAmelCase : Dict =seq_length UpperCAmelCase : Optional[Any] =is_training UpperCAmelCase : List[str] =use_labels UpperCAmelCase : List[Any] =vocab_size UpperCAmelCase : Optional[int] =hidden_size UpperCAmelCase : Tuple =num_hidden_layers UpperCAmelCase : Any =num_attention_heads UpperCAmelCase : Optional[int] =intermediate_size UpperCAmelCase : str =hidden_dropout_prob UpperCAmelCase : Optional[int] =attention_probs_dropout_prob UpperCAmelCase : str =max_position_embeddings UpperCAmelCase : List[Any] =eos_token_id UpperCAmelCase : Optional[int] =pad_token_id UpperCAmelCase : Tuple =bos_token_id def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : List[Any] =ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase : List[Any] =tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase : Tuple =tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase : str =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[Any] =self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase : List[str] =prepare_blenderbot_inputs_dict(snake_case__ , snake_case__ , snake_case__ ) return config, inputs_dict def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> int: '''simple docstring''' UpperCAmelCase : Union[str, Any] =TFBlenderbotModel(config=snake_case__ ).get_decoder() UpperCAmelCase : Any =inputs_dict['''input_ids'''] UpperCAmelCase : str =input_ids[:1, :] UpperCAmelCase : Tuple =inputs_dict['''attention_mask'''][:1, :] UpperCAmelCase : Tuple =inputs_dict['''head_mask'''] UpperCAmelCase : List[Any] =1 # first forward pass UpperCAmelCase : List[str] =model(snake_case__ , attention_mask=snake_case__ , head_mask=snake_case__ , use_cache=snake_case__ ) UpperCAmelCase , UpperCAmelCase : str =outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase : Union[str, Any] =ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase : List[Any] =tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase : Tuple =tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase : int =tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase : Optional[int] =model(snake_case__ , attention_mask=snake_case__ )[0] UpperCAmelCase : str =model(snake_case__ , attention_mask=snake_case__ , past_key_values=snake_case__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase : List[Any] =int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase : List[Any] =output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase : Dict =output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(snake_case__ , snake_case__ , rtol=1e-3 ) def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , )-> str: '''simple docstring''' if attention_mask is None: UpperCAmelCase : int =tf.cast(tf.math.not_equal(__lowerCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase : Tuple =tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase : str =tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase : Union[str, Any] =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase : int =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __snake_case ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase : List[str] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __lowerCamelCase : Dict = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __lowerCamelCase : Dict = ( { """conversational""": TFBlenderbotForConditionalGeneration, """feature-extraction""": TFBlenderbotModel, """summarization""": TFBlenderbotForConditionalGeneration, """text2text-generation""": TFBlenderbotForConditionalGeneration, """translation""": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __lowerCamelCase : Union[str, Any] = True __lowerCamelCase : Union[str, Any] = False __lowerCamelCase : Union[str, Any] = False def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : List[str] =TFBlenderbotModelTester(self ) UpperCAmelCase : List[Any] =ConfigTester(self , config_class=snake_case__ ) def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : int =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*snake_case__ ) @require_tokenizers @require_tf class __snake_case ( unittest.TestCase ): __lowerCamelCase : List[str] = ["""My friends are cool but they eat too many carbs."""] __lowerCamelCase : Dict = """facebook/blenderbot-400M-distill""" @cached_property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' UpperCAmelCase : int =TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Optional[int] =self.tokenizer(self.src_text , return_tensors='''tf''' ) UpperCAmelCase : Optional[int] =self.model.generate( model_inputs.input_ids , ) UpperCAmelCase : str =self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=snake_case__ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
348
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def __lowerCamelCase ( snake_case__ ,snake_case__=False ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("""head""" ): _SCREAMING_SNAKE_CASE = """segformer.encoder.""" + key if key.startswith("""backbone""" ): _SCREAMING_SNAKE_CASE = key.replace("""backbone""" ,"""segformer.encoder""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _SCREAMING_SNAKE_CASE = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] _SCREAMING_SNAKE_CASE = key.replace(F'patch_embed{idx}' ,F'patch_embeddings.{int(snake_case__ )-1}' ) if "norm" in key: _SCREAMING_SNAKE_CASE = key.replace("""norm""" ,"""layer_norm""" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _SCREAMING_SNAKE_CASE = key[key.find("""segformer.encoder.layer_norm""" ) + len("""segformer.encoder.layer_norm""" )] _SCREAMING_SNAKE_CASE = key.replace(F'layer_norm{idx}' ,F'layer_norm.{int(snake_case__ )-1}' ) if "layer_norm1" in key: _SCREAMING_SNAKE_CASE = key.replace("""layer_norm1""" ,"""layer_norm_1""" ) if "layer_norm2" in key: _SCREAMING_SNAKE_CASE = key.replace("""layer_norm2""" ,"""layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 _SCREAMING_SNAKE_CASE = key[key.find("""block""" ) + len("""block""" )] _SCREAMING_SNAKE_CASE = key.replace(F'block{idx}' ,F'block.{int(snake_case__ )-1}' ) if "attn.q" in key: _SCREAMING_SNAKE_CASE = key.replace("""attn.q""" ,"""attention.self.query""" ) if "attn.proj" in key: _SCREAMING_SNAKE_CASE = key.replace("""attn.proj""" ,"""attention.output.dense""" ) if "attn" in key: _SCREAMING_SNAKE_CASE = key.replace("""attn""" ,"""attention.self""" ) if "fc1" in key: _SCREAMING_SNAKE_CASE = key.replace("""fc1""" ,"""dense1""" ) if "fc2" in key: _SCREAMING_SNAKE_CASE = key.replace("""fc2""" ,"""dense2""" ) if "linear_pred" in key: _SCREAMING_SNAKE_CASE = key.replace("""linear_pred""" ,"""classifier""" ) if "linear_fuse" in key: _SCREAMING_SNAKE_CASE = key.replace("""linear_fuse.conv""" ,"""linear_fuse""" ) _SCREAMING_SNAKE_CASE = key.replace("""linear_fuse.bn""" ,"""batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _SCREAMING_SNAKE_CASE = key[key.find("""linear_c""" ) + len("""linear_c""" )] _SCREAMING_SNAKE_CASE = key.replace(F'linear_c{idx}' ,F'linear_c.{int(snake_case__ )-1}' ) if key.startswith("""head""" ): _SCREAMING_SNAKE_CASE = key.replace("""head""" ,"""classifier""" ) _SCREAMING_SNAKE_CASE = value return new_state_dict def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Dict: """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _SCREAMING_SNAKE_CASE = state_dict.pop(F'segformer.encoder.block.{i}.{j}.attention.self.kv.weight' ) _SCREAMING_SNAKE_CASE = state_dict.pop(F'segformer.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict _SCREAMING_SNAKE_CASE = kv_weight[ : config.hidden_sizes[i], : ] _SCREAMING_SNAKE_CASE = kv_bias[: config.hidden_sizes[i]] _SCREAMING_SNAKE_CASE = kv_weight[ config.hidden_sizes[i] :, : ] _SCREAMING_SNAKE_CASE = kv_bias[ config.hidden_sizes[i] : ] def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = """http://images.cocodataset.org/val2017/000000039769.jpg""" _SCREAMING_SNAKE_CASE = Image.open(requests.get(snake_case__ ,stream=snake_case__ ).raw ) return image @torch.no_grad() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = SegformerConfig() _SCREAMING_SNAKE_CASE = False # set attributes based on model_name _SCREAMING_SNAKE_CASE = """huggingface/label-files""" if "segformer" in model_name: _SCREAMING_SNAKE_CASE = model_name[len("""segformer.""" ) : len("""segformer.""" ) + 2] if "ade" in model_name: _SCREAMING_SNAKE_CASE = 1_50 _SCREAMING_SNAKE_CASE = """ade20k-id2label.json""" _SCREAMING_SNAKE_CASE = (1, 1_50, 1_28, 1_28) elif "city" in model_name: _SCREAMING_SNAKE_CASE = 19 _SCREAMING_SNAKE_CASE = """cityscapes-id2label.json""" _SCREAMING_SNAKE_CASE = (1, 19, 1_28, 1_28) else: raise ValueError(F'Model {model_name} not supported' ) elif "mit" in model_name: _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = model_name[4:6] _SCREAMING_SNAKE_CASE = 10_00 _SCREAMING_SNAKE_CASE = """imagenet-1k-id2label.json""" _SCREAMING_SNAKE_CASE = (1, 10_00) else: raise ValueError(F'Model {model_name} not supported' ) # set config attributes _SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(snake_case__ ,snake_case__ ,repo_type="""dataset""" ) ,"""r""" ) ) _SCREAMING_SNAKE_CASE = {int(snake_case__ ): v for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE = idalabel _SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _SCREAMING_SNAKE_CASE = [64, 1_28, 3_20, 5_12] _SCREAMING_SNAKE_CASE = 2_56 elif size == "b2": _SCREAMING_SNAKE_CASE = [64, 1_28, 3_20, 5_12] _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = [3, 4, 6, 3] elif size == "b3": _SCREAMING_SNAKE_CASE = [64, 1_28, 3_20, 5_12] _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = [3, 4, 18, 3] elif size == "b4": _SCREAMING_SNAKE_CASE = [64, 1_28, 3_20, 5_12] _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = [3, 8, 27, 3] elif size == "b5": _SCREAMING_SNAKE_CASE = [64, 1_28, 3_20, 5_12] _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = [3, 6, 40, 3] else: raise ValueError(F'Size {size} not supported' ) # load image processor (only resize + normalize) _SCREAMING_SNAKE_CASE = SegformerImageProcessor( image_scale=(5_12, 5_12) ,keep_ratio=snake_case__ ,align=snake_case__ ,do_random_crop=snake_case__ ) # prepare image _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=snake_case__ ,return_tensors="""pt""" ).pixel_values logger.info(F'Converting model {model_name}...' ) # load original state dict if encoder_only: _SCREAMING_SNAKE_CASE = torch.load(snake_case__ ,map_location=torch.device("""cpu""" ) ) else: _SCREAMING_SNAKE_CASE = torch.load(snake_case__ ,map_location=torch.device("""cpu""" ) )["""state_dict"""] # rename keys _SCREAMING_SNAKE_CASE = rename_keys(snake_case__ ,encoder_only=snake_case__ ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(snake_case__ ,snake_case__ ) # create HuggingFace model and load state dict if encoder_only: _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = SegformerForImageClassification(snake_case__ ) else: _SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation(snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() # forward pass _SCREAMING_SNAKE_CASE = model(snake_case__ ) _SCREAMING_SNAKE_CASE = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-4.6_310, -5.5_232, -6.2_356], [-5.1_921, -6.1_444, -6.5_996], [-5.4_424, -6.2_790, -6.7_574]], [[-12.1_391, -13.3_122, -13.9_554], [-12.8_732, -13.9_352, -14.3_563], [-12.9_438, -13.8_226, -14.2_513]], [[-12.5_134, -13.4_686, -14.4_915], [-12.8_669, -14.4_343, -14.7_758], [-13.2_523, -14.5_819, -15.0_694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-7.5_820, -8.7_231, -8.3_215], [-8.0_600, -10.3_529, -10.0_304], [-7.5_208, -9.4_103, -9.6_239]], [[-12.6_918, -13.8_994, -13.7_137], [-13.3_196, -15.7_523, -15.4_789], [-12.9_343, -14.8_757, -14.9_689]], [[-11.1_911, -11.9_421, -11.3_243], [-11.3_342, -13.6_839, -13.3_581], [-10.3_909, -12.1_832, -12.4_858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-11.8_173, -14.3_850, -16.3_128], [-14.5_648, -16.5_804, -18.6_568], [-14.7_223, -15.7_387, -18.4_218]], [[-15.7_290, -17.9_171, -19.4_423], [-18.3_105, -19.9_448, -21.4_661], [-17.9_296, -18.6_497, -20.7_910]], [[-15.0_783, -17.0_336, -18.2_789], [-16.8_771, -18.6_870, -20.1_612], [-16.2_454, -17.1_426, -19.5_055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-9.0_878, -10.2_081, -10.1_891], [-9.3_144, -10.7_941, -10.9_843], [-9.2_294, -10.3_855, -10.5_704]], [[-12.2_316, -13.9_068, -13.6_102], [-12.9_161, -14.3_702, -14.3_235], [-12.5_233, -13.7_174, -13.7_932]], [[-14.6_275, -15.2_490, -14.9_727], [-14.3_400, -15.9_687, -16.2_827], [-14.1_484, -15.4_033, -15.8_937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-12.3_144, -13.2_447, -14.0_802], [-13.3_614, -14.5_816, -15.6_117], [-13.3_340, -14.4_433, -16.2_219]], [[-19.2_781, -20.4_128, -20.7_506], [-20.6_153, -21.6_566, -22.0_998], [-19.9_800, -21.0_430, -22.1_494]], [[-18.8_739, -19.7_804, -21.1_834], [-20.1_233, -21.6_765, -23.2_944], [-20.0_315, -21.2_641, -23.6_944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-9.5_524, -12.0_835, -11.7_348], [-10.5_229, -13.6_446, -14.5_662], [-9.5_842, -12.8_851, -13.9_414]], [[-15.3_432, -17.5_323, -17.0_818], [-16.3_330, -18.9_255, -19.2_101], [-15.1_340, -17.7_848, -18.3_971]], [[-12.6_072, -14.9_486, -14.6_631], [-13.7_629, -17.0_907, -17.7_745], [-12.7_899, -16.1_695, -17.1_671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-11.9_295, -13.4_057, -14.8_106], [-13.3_431, -14.8_179, -15.3_781], [-14.2_836, -15.5_942, -16.1_588]], [[-11.4_906, -12.8_067, -13.6_564], [-13.1_189, -14.0_500, -14.1_543], [-13.8_748, -14.5_136, -14.8_789]], [[0.5_374, 0.1_067, -0.4_742], [0.1_141, -0.2_255, -0.7_099], [-0.3_000, -0.5_924, -1.3_105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-7.8_217, -9.8_767, -10.1_717], [-9.4_438, -10.9_058, -11.4_047], [-9.7_939, -12.3_495, -12.1_079]], [[-7.1_514, -9.5_336, -10.0_860], [-9.7_776, -11.6_822, -11.8_439], [-10.1_411, -12.7_655, -12.8_972]], [[0.3_021, 0.0_805, -0.2_310], [-0.0_328, -0.1_605, -0.2_714], [-0.1_408, -0.5_477, -0.6_976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [ [-1.1_3_7_2e0_1, -1.2_7_8_7e0_1, -1.3_4_7_7e0_1], [-1.2_5_3_6e0_1, -1.4_1_9_4e0_1, -1.4_4_0_9e0_1], [-1.3_2_1_7e0_1, -1.4_8_8_8e0_1, -1.5_3_2_7e0_1], ], [ [-1.4_7_9_1e0_1, -1.7_1_2_2e0_1, -1.8_2_7_7e0_1], [-1.7_1_6_3e0_1, -1.9_1_9_2e0_1, -1.9_5_3_3e0_1], [-1.7_8_9_7e0_1, -1.9_9_9_1e0_1, -2.0_3_1_5e0_1], ], [ [7.6_7_2_3e-0_1, 4.1_9_2_1e-0_1, -7.7_8_7_8e-0_2], [4.7_7_7_2e-0_1, 9.5_5_5_7e-0_3, -2.8_0_8_2e-0_1], [3.6_0_3_2e-0_1, -2.4_8_2_6e-0_1, -5.1_1_6_8e-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-9.4_959, -11.3_087, -11.7_479], [-11.0_025, -12.6_540, -12.3_319], [-11.4_064, -13.0_487, -12.9_905]], [[-9.8_905, -11.3_084, -12.0_854], [-11.1_726, -12.7_698, -12.9_583], [-11.5_985, -13.3_278, -14.1_774]], [[0.2_213, 0.0_192, -0.2_466], [-0.1_731, -0.4_213, -0.4_874], [-0.3_126, -0.6_541, -1.1_389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-13.5_748, -13.9_111, -12.6_500], [-14.3_500, -15.3_683, -14.2_328], [-14.7_532, -16.0_424, -15.6_087]], [[-17.1_651, -15.8_725, -12.9_653], [-17.2_580, -17.3_718, -14.8_223], [-16.6_058, -16.8_783, -16.7_452]], [[-3.6_456, -3.0_209, -1.4_203], [-3.0_797, -3.1_959, -2.0_000], [-1.8_757, -1.9_217, -1.6_997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-16.0_976, -16.4_856, -17.3_962], [-16.6_234, -19.0_342, -19.7_685], [-16.0_900, -18.0_661, -19.1_180]], [[-18.4_750, -18.8_488, -19.5_074], [-19.4_030, -22.1_570, -22.5_977], [-19.1_191, -20.8_486, -22.3_783]], [[-4.5_178, -5.5_037, -6.5_109], [-5.0_884, -7.2_174, -8.0_334], [-4.4_156, -5.8_117, -7.2_970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-14.2_081, -14.4_732, -14.1_977], [-14.5_867, -16.4_423, -16.6_356], [-13.4_441, -14.9_685, -16.8_696]], [[-14.4_576, -14.7_073, -15.0_451], [-15.0_816, -17.6_237, -17.9_873], [-14.4_213, -16.0_199, -18.5_992]], [[-4.7_349, -4.9_588, -5.0_966], [-4.3_210, -6.9_325, -7.2_591], [-3.4_312, -4.7_484, -7.1_917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-11.7_737, -11.9_526, -11.3_273], [-13.6_692, -14.4_574, -13.8_878], [-13.8_937, -14.6_924, -15.9_345]], [[-14.6_706, -14.5_330, -14.1_306], [-16.1_502, -16.8_180, -16.4_269], [-16.8_338, -17.8_939, -20.1_746]], [[1.0_491, 0.8_289, 1.0_310], [1.1_044, 0.5_219, 0.8_055], [1.0_899, 0.6_926, 0.5_590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _SCREAMING_SNAKE_CASE = torch.tensor( [ [[-12.5_641, -13.4_777, -13.0_684], [-13.9_587, -15.8_983, -16.6_557], [-13.3_109, -15.7_350, -16.3_141]], [[-14.7_074, -15.4_352, -14.5_944], [-16.6_353, -18.1_663, -18.6_120], [-15.1_702, -18.0_329, -18.1_547]], [[-1.7_990, -2.0_951, -1.7_784], [-2.6_397, -3.8_245, -3.9_686], [-1.5_264, -2.8_126, -2.9_316]], ] ) else: _SCREAMING_SNAKE_CASE = logits.argmax(-1 ).item() print("""Predicted class:""" ,model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] ,snake_case__ ,atol=1e-2 ) # finally, save model and image processor logger.info(F'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
125
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCamelCase = { '''configuration_layoutlmv3''': [ '''LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LayoutLMv3Config''', '''LayoutLMv3OnnxConfig''', ], '''processing_layoutlmv3''': ['''LayoutLMv3Processor'''], '''tokenization_layoutlmv3''': ['''LayoutLMv3Tokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ['''LayoutLMv3TokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LayoutLMv3ForQuestionAnswering''', '''LayoutLMv3ForSequenceClassification''', '''LayoutLMv3ForTokenClassification''', '''LayoutLMv3Model''', '''LayoutLMv3PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLayoutLMv3ForQuestionAnswering''', '''TFLayoutLMv3ForSequenceClassification''', '''TFLayoutLMv3ForTokenClassification''', '''TFLayoutLMv3Model''', '''TFLayoutLMv3PreTrainedModel''', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ['''LayoutLMv3FeatureExtractor'''] UpperCamelCase = ['''LayoutLMv3ImageProcessor'''] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
125
1
def _A ( _lowercase , _lowercase ) -> List[str]: """simple docstring""" return number | (1 << position) def _A ( _lowercase , _lowercase ) -> Union[str, Any]: """simple docstring""" return number & ~(1 << position) def _A ( _lowercase , _lowercase ) -> List[str]: """simple docstring""" return number ^ (1 << position) def _A ( _lowercase , _lowercase ) -> Tuple: """simple docstring""" return ((number >> position) & 1) == 1 def _A ( _lowercase , _lowercase ) -> Dict: """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
310
def a ( snake_case__: list ): '''simple docstring''' if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] lowercase_ = [] def generate(snake_case__: int , snake_case__: list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even lowercase_ , lowercase_ = arr[k - 1], arr[i] else: # k is odd lowercase_ , lowercase_ = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": __a = input('Enter numbers separated by a comma:\n').strip() __a = [int(item) for item in user_input.split(',')] print(heaps(arr))
30
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''xlm-mlm-en-2048''': '''https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json''', '''xlm-mlm-ende-1024''': '''https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json''', '''xlm-mlm-enfr-1024''': '''https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json''', '''xlm-mlm-enro-1024''': '''https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json''', '''xlm-mlm-tlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json''', '''xlm-mlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json''', '''xlm-clm-enfr-1024''': '''https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json''', '''xlm-clm-ende-1024''': '''https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json''', '''xlm-mlm-17-1280''': '''https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json''', '''xlm-mlm-100-1280''': '''https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json''', } class __magic_name__ (__SCREAMING_SNAKE_CASE ): lowerCamelCase__ = 'xlm' lowerCamelCase__ = { 'hidden_size': 'emb_dim', 'num_attention_heads': 'n_heads', 'num_hidden_layers': 'n_layers', 'n_words': 'vocab_size', # For backward compatibility } def __init__( self , _a=30145 , _a=2048 , _a=12 , _a=16 , _a=0.1 , _a=0.1 , _a=True , _a=False , _a=False , _a=False , _a=1 , _a=True , _a=512 , _a=2048**-0.5 , _a=1E-12 , _a=0.0_2 , _a=0 , _a=1 , _a=2 , _a=3 , _a=5 , _a=True , _a="first" , _a=True , _a=None , _a=True , _a=0.1 , _a=5 , _a=5 , _a=0 , _a=0 , _a=2 , _a=0 , **_a , ) -> List[Any]: lowerCAmelCase_ = vocab_size lowerCAmelCase_ = emb_dim lowerCAmelCase_ = n_layers lowerCAmelCase_ = n_heads lowerCAmelCase_ = dropout lowerCAmelCase_ = attention_dropout lowerCAmelCase_ = gelu_activation lowerCAmelCase_ = sinusoidal_embeddings lowerCAmelCase_ = causal lowerCAmelCase_ = asm lowerCAmelCase_ = n_langs lowerCAmelCase_ = use_lang_emb lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = bos_index lowerCAmelCase_ = eos_index lowerCAmelCase_ = pad_index lowerCAmelCase_ = unk_index lowerCAmelCase_ = mask_index lowerCAmelCase_ = is_encoder lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = embed_init_std lowerCAmelCase_ = init_std lowerCAmelCase_ = summary_type lowerCAmelCase_ = summary_use_proj lowerCAmelCase_ = summary_activation lowerCAmelCase_ = summary_proj_to_labels lowerCAmelCase_ = summary_first_dropout lowerCAmelCase_ = start_n_top lowerCAmelCase_ = end_n_top lowerCAmelCase_ = mask_token_id lowerCAmelCase_ = lang_id if "n_words" in kwargs: lowerCAmelCase_ = kwargs["n_words"] super().__init__(pad_token_id=_a , bos_token_id=_a , **_a ) class __magic_name__ (__SCREAMING_SNAKE_CASE ): @property def __a ( self ) -> Dict: if self.task == "multiple-choice": lowerCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: lowerCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
361
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def A(__a: Tuple , __a: Union[str, Any] ): lowerCAmelCase_ = checkpoint lowerCAmelCase_ = {} lowerCAmelCase_ = vae_state_dict["encoder.conv_in.weight"] lowerCAmelCase_ = vae_state_dict["encoder.conv_in.bias"] lowerCAmelCase_ = vae_state_dict["encoder.conv_out.weight"] lowerCAmelCase_ = vae_state_dict["encoder.conv_out.bias"] lowerCAmelCase_ = vae_state_dict["encoder.norm_out.weight"] lowerCAmelCase_ = vae_state_dict["encoder.norm_out.bias"] lowerCAmelCase_ = vae_state_dict["decoder.conv_in.weight"] lowerCAmelCase_ = vae_state_dict["decoder.conv_in.bias"] lowerCAmelCase_ = vae_state_dict["decoder.conv_out.weight"] lowerCAmelCase_ = vae_state_dict["decoder.conv_out.bias"] lowerCAmelCase_ = vae_state_dict["decoder.norm_out.weight"] lowerCAmelCase_ = vae_state_dict["decoder.norm_out.bias"] lowerCAmelCase_ = vae_state_dict["quant_conv.weight"] lowerCAmelCase_ = vae_state_dict["quant_conv.bias"] lowerCAmelCase_ = vae_state_dict["post_quant_conv.weight"] lowerCAmelCase_ = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only lowerCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) lowerCAmelCase_ = { layer_id: [key for key in vae_state_dict if F"down.{layer_id}" in key] for layer_id in range(__a ) } # Retrieves the keys for the decoder up blocks only lowerCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) lowerCAmelCase_ = { layer_id: [key for key in vae_state_dict if F"up.{layer_id}" in key] for layer_id in range(__a ) } for i in range(__a ): lowerCAmelCase_ = [key for key in down_blocks[i] if F"down.{i}" in key and F"down.{i}.downsample" not in key] if F"encoder.down.{i}.downsample.conv.weight" in vae_state_dict: lowerCAmelCase_ = vae_state_dict.pop( F"encoder.down.{i}.downsample.conv.weight" ) lowerCAmelCase_ = vae_state_dict.pop( F"encoder.down.{i}.downsample.conv.bias" ) lowerCAmelCase_ = renew_vae_resnet_paths(__a ) lowerCAmelCase_ = {"old": F"down.{i}.block", "new": F"down_blocks.{i}.resnets"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) lowerCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.block" in key] lowerCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): lowerCAmelCase_ = [key for key in mid_resnets if F"encoder.mid.block_{i}" in key] lowerCAmelCase_ = renew_vae_resnet_paths(__a ) lowerCAmelCase_ = {"old": F"mid.block_{i}", "new": F"mid_block.resnets.{i - 1}"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) lowerCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.attn" in key] lowerCAmelCase_ = renew_vae_attention_paths(__a ) lowerCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) conv_attn_to_linear(__a ) for i in range(__a ): lowerCAmelCase_ = num_up_blocks - 1 - i lowerCAmelCase_ = [ key for key in up_blocks[block_id] if F"up.{block_id}" in key and F"up.{block_id}.upsample" not in key ] if F"decoder.up.{block_id}.upsample.conv.weight" in vae_state_dict: lowerCAmelCase_ = vae_state_dict[ F"decoder.up.{block_id}.upsample.conv.weight" ] lowerCAmelCase_ = vae_state_dict[ F"decoder.up.{block_id}.upsample.conv.bias" ] lowerCAmelCase_ = renew_vae_resnet_paths(__a ) lowerCAmelCase_ = {"old": F"up.{block_id}.block", "new": F"up_blocks.{i}.resnets"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) lowerCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.block" in key] lowerCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): lowerCAmelCase_ = [key for key in mid_resnets if F"decoder.mid.block_{i}" in key] lowerCAmelCase_ = renew_vae_resnet_paths(__a ) lowerCAmelCase_ = {"old": F"mid.block_{i}", "new": F"mid_block.resnets.{i - 1}"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) lowerCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.attn" in key] lowerCAmelCase_ = renew_vae_attention_paths(__a ) lowerCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__a , __a , __a , additional_replacements=[meta_path] , config=__a ) conv_attn_to_linear(__a ) return new_checkpoint def A(__a: str , __a: str , ): # Only support V1 lowerCAmelCase_ = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) lowerCAmelCase_ = io.BytesIO(r.content ) lowerCAmelCase_ = OmegaConf.load(__a ) lowerCAmelCase_ = 512 lowerCAmelCase_ = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open lowerCAmelCase_ = {} with safe_open(__a , framework="pt" , device="cpu" ) as f: for key in f.keys(): lowerCAmelCase_ = f.get_tensor(__a ) else: lowerCAmelCase_ = torch.load(__a , map_location=__a )["state_dict"] # Convert the VAE model. lowerCAmelCase_ = create_vae_diffusers_config(__a , image_size=__a ) lowerCAmelCase_ = custom_convert_ldm_vae_checkpoint(__a , __a ) lowerCAmelCase_ = AutoencoderKL(**__a ) vae.load_state_dict(__a ) vae.save_pretrained(__a ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument('''--vae_pt_path''', default=None, type=str, required=True, help='''Path to the VAE.pt to convert.''') parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the VAE.pt to convert.''') lowerCamelCase__ = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
22
0
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __snake_case ( UpperCamelCase_ ,UpperCamelCase_ ,unittest.TestCase ): _a = IFInpaintingPipeline _a = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _a = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _a = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCAmelCase__ ( self : Optional[int]): return self._get_dummy_components() def UpperCAmelCase__ ( self : List[str] , A_ : Dict , A_ : str=0): if str(A_).startswith('''mps'''): lowerCAmelCase_ : Dict = torch.manual_seed(A_) else: lowerCAmelCase_ : int = torch.Generator(device=A_).manual_seed(A_) lowerCAmelCase_ : Optional[int] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(A_)).to(A_) lowerCAmelCase_ : List[str] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(A_)).to(A_) lowerCAmelCase_ : int = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCAmelCase__ ( self : Any): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3) def UpperCAmelCase__ ( self : Union[str, Any]): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''') def UpperCAmelCase__ ( self : List[str]): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1) def UpperCAmelCase__ ( self : List[Any]): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2) def UpperCAmelCase__ ( self : Any): self._test_save_load_local() def UpperCAmelCase__ ( self : List[str]): self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
103
from pathlib import Path import fire def UpperCamelCase( __UpperCamelCase : str ,__UpperCamelCase : str ,__UpperCamelCase : int ): lowerCAmelCase_ : List[str] = Path(__UpperCamelCase ) lowerCAmelCase_ : Union[str, Any] = Path(__UpperCamelCase ) dest_dir.mkdir(exist_ok=__UpperCamelCase ) for path in src_dir.iterdir(): lowerCAmelCase_ : Optional[Any] = [x.rstrip() for x in list(path.open().readlines() )][:n] lowerCAmelCase_ : List[str] = dest_dir.joinpath(path.name ) print(__UpperCamelCase ) dest_path.open('''w''' ).write('''\n'''.join(__UpperCamelCase ) ) if __name__ == "__main__": fire.Fire(minify)
103
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Tuple = logging.get_logger(__name__) _lowerCamelCase : Tuple = { '''weiweishi/roc-bert-base-zh''': '''https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json''', } class lowercase ( a ): lowercase__ : List[str] = """roc_bert""" def __init__( self : Dict , _UpperCamelCase : Dict=30_522 , _UpperCamelCase : Tuple=768 , _UpperCamelCase : int=12 , _UpperCamelCase : str=12 , _UpperCamelCase : Any=3_072 , _UpperCamelCase : Tuple="gelu" , _UpperCamelCase : int=0.1 , _UpperCamelCase : List[str]=0.1 , _UpperCamelCase : Tuple=512 , _UpperCamelCase : Tuple=2 , _UpperCamelCase : Dict=0.0_2 , _UpperCamelCase : Dict=1e-12 , _UpperCamelCase : Optional[int]=True , _UpperCamelCase : int=0 , _UpperCamelCase : Tuple="absolute" , _UpperCamelCase : Any=None , _UpperCamelCase : List[str]=True , _UpperCamelCase : Any=True , _UpperCamelCase : Any=768 , _UpperCamelCase : List[Any]=910 , _UpperCamelCase : List[str]=512 , _UpperCamelCase : Union[str, Any]=24_858 , _UpperCamelCase : List[str]=True , **_UpperCamelCase : int , ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = type_vocab_size SCREAMING_SNAKE_CASE = layer_norm_eps SCREAMING_SNAKE_CASE = use_cache SCREAMING_SNAKE_CASE = enable_pronunciation SCREAMING_SNAKE_CASE = enable_shape SCREAMING_SNAKE_CASE = pronunciation_embed_dim SCREAMING_SNAKE_CASE = pronunciation_vocab_size SCREAMING_SNAKE_CASE = shape_embed_dim SCREAMING_SNAKE_CASE = shape_vocab_size SCREAMING_SNAKE_CASE = concat_input SCREAMING_SNAKE_CASE = position_embedding_type SCREAMING_SNAKE_CASE = classifier_dropout super().__init__(pad_token_id=_UpperCamelCase , **_UpperCamelCase )
206
# Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar _lowerCamelCase : Union[str, Any] = TypeVar('''T''') class lowercase ( Generic[T] ): def __init__( self : int , _UpperCamelCase : bool = True ) -> None: '''simple docstring''' SCREAMING_SNAKE_CASE = {} # dictionary of lists SCREAMING_SNAKE_CASE = directed def __snake_case( self : int , _UpperCamelCase : T , _UpperCamelCase : T ) -> GraphAdjacencyList[T]: '''simple docstring''' if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(_UpperCamelCase ) self.adj_list[destination_vertex].append(_UpperCamelCase ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(_UpperCamelCase ) SCREAMING_SNAKE_CASE = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(_UpperCamelCase ) SCREAMING_SNAKE_CASE = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: SCREAMING_SNAKE_CASE = [destination_vertex] SCREAMING_SNAKE_CASE = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(_UpperCamelCase ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(_UpperCamelCase ) SCREAMING_SNAKE_CASE = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: SCREAMING_SNAKE_CASE = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: SCREAMING_SNAKE_CASE = [destination_vertex] SCREAMING_SNAKE_CASE = [] return self def __repr__( self : Union[str, Any] ) -> str: '''simple docstring''' return pformat(self.adj_list )
206
1
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Tuple = inspect.getfile(accelerate.test_utils ) _snake_case : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) _snake_case : Optional[Any] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) _snake_case : int = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices." ) _snake_case : Optional[int] = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: Dict ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices." ) _snake_case : Optional[int] = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.operation_file_path] print(f"Command: {cmd}" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : int = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices, using 2 devices only" ) _snake_case : Tuple = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.data_loop_file_path] with patch_environment(omp_num_threads=1, cuda_visible_devices="""0,1""" ): execute_subprocess_async(a_, env=os.environ.copy() ) if __name__ == "__main__": A_ = Accelerator() A_ = (accelerator.state.process_index + 2, 10) A_ = torch.randint(0, 10, shape).to(accelerator.device) A_ = '''''' A_ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." A_ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." A_ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
64
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Any = logging.get_logger(__name__) __A : Dict = {'vocab_file': 'spiece.model'} __A : List[Any] = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<sep>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE="<cls>" , _SCREAMING_SNAKE_CASE="<mask>" , _SCREAMING_SNAKE_CASE=["<eop>", "<eod>"] , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , )-> None: lowerCamelCase_ =AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else mask_token lowerCamelCase_ ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_SCREAMING_SNAKE_CASE , remove_space=_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **_SCREAMING_SNAKE_CASE , ) lowerCamelCase_ =3 lowerCamelCase_ =do_lower_case lowerCamelCase_ =remove_space lowerCamelCase_ =keep_accents lowerCamelCase_ =vocab_file lowerCamelCase_ =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_SCREAMING_SNAKE_CASE ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) lowerCamelCase_ =jieba lowerCamelCase_ =str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _snake_case ( self )-> Any: return len(self.sp_model ) def _snake_case ( self )-> Dict: lowerCamelCase_ ={self.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self )-> List[Any]: lowerCamelCase_ =self.__dict__.copy() lowerCamelCase_ =None return state def __setstate__( self , _SCREAMING_SNAKE_CASE )-> List[Any]: lowerCamelCase_ =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase_ ={} lowerCamelCase_ =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> int: if self.remove_space: lowerCamelCase_ =""" """.join(inputs.strip().split() ) else: lowerCamelCase_ =inputs lowerCamelCase_ =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: lowerCamelCase_ =unicodedata.normalize("""NFKD""" , _SCREAMING_SNAKE_CASE ) lowerCamelCase_ ="""""".join([c for c in outputs if not unicodedata.combining(_SCREAMING_SNAKE_CASE )] ) if self.do_lower_case: lowerCamelCase_ =outputs.lower() return outputs def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> List[str]: lowerCamelCase_ =self.preprocess_text(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =self.sp_model.encode(_SCREAMING_SNAKE_CASE , out_type=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =[] for piece in pieces: if len(_SCREAMING_SNAKE_CASE ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowerCamelCase_ =self.sp_model.EncodeAsPieces(piece[:-1].replace(_SCREAMING_SNAKE_CASE , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase_ =cur_pieces[1:] else: lowerCamelCase_ =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_SCREAMING_SNAKE_CASE ) else: new_pieces.append(_SCREAMING_SNAKE_CASE ) return new_pieces def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: return self.sp_model.PieceToId(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: return self.sp_model.IdToPiece(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: lowerCamelCase_ ="""""".join(_SCREAMING_SNAKE_CASE ).replace(_SCREAMING_SNAKE_CASE , """ """ ).strip() return out_string def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> List[int]: lowerCamelCase_ =[self.sep_token_id] lowerCamelCase_ =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False )-> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_SCREAMING_SNAKE_CASE , token_ids_a=_SCREAMING_SNAKE_CASE , already_has_special_tokens=_SCREAMING_SNAKE_CASE ) if token_ids_a is not None: return ([0] * len(_SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(_SCREAMING_SNAKE_CASE )) + [1, 1] return ([0] * len(_SCREAMING_SNAKE_CASE )) + [1, 1] def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> List[int]: lowerCamelCase_ =[self.sep_token_id] lowerCamelCase_ =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> Tuple[str]: if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowerCamelCase_ =os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(_SCREAMING_SNAKE_CASE , """wb""" ) as fi: lowerCamelCase_ =self.sp_model.serialized_model_proto() fi.write(_SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _snake_case ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )-> Any: lowerCamelCase_ =super()._decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
154
0
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def snake_case__ ( SCREAMING_SNAKE_CASE_ : BertModel , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str ): '''simple docstring''' lowercase__ : Tuple = ('dense.weight', 'attention.self.query', 'attention.self.key', 'attention.self.value') lowercase__ : Dict = ( ('layer.', 'layer_'), ('word_embeddings.weight', 'word_embeddings'), ('position_embeddings.weight', 'position_embeddings'), ('token_type_embeddings.weight', 'token_type_embeddings'), ('.', '/'), ('LayerNorm/weight', 'LayerNorm/gamma'), ('LayerNorm/bias', 'LayerNorm/beta'), ('weight', 'kernel'), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ : str = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ : str ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ : Optional[int] = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f"""bert/{name}""" def create_tf_var(SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : tf.Session ): lowercase__ : Optional[Any] = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ : List[str] = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ : Tuple = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ : str = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ : str = torch_tensor.T lowercase__ : Any = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ : Optional[int] = session.run(SCREAMING_SNAKE_CASE_ ) print(f"""Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}""" ) lowercase__ : Tuple = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace('-' , '_' ) + '.ckpt' ) ) def snake_case__ ( SCREAMING_SNAKE_CASE_ : Optional[Any]=None ): '''simple docstring''' lowercase__ : List[str] = argparse.ArgumentParser() parser.add_argument('--model_name' , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help='model name e.g. bert-base-uncased' ) parser.add_argument( '--cache_dir' , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help='Directory containing pytorch model' ) parser.add_argument('--pytorch_model_path' , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help='/path/to/<pytorch-model-name>.bin' ) parser.add_argument('--tf_cache_dir' , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help='Directory in which to save tensorflow model' ) lowercase__ : List[str] = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ : int = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
359
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore snake_case_ = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" snake_case_ = [file for file in filepaths if file != file.lower()] if upper_files: print(F'''{len(upper_files)} files contain uppercase characters:''') print('''\n'''.join(upper_files) + '''\n''') snake_case_ = [file for file in filepaths if ''' ''' in file] if space_files: print(F'''{len(space_files)} files contain space characters:''') print('''\n'''.join(space_files) + '''\n''') snake_case_ = [file for file in filepaths if '''-''' in file] if hyphen_files: print(F'''{len(hyphen_files)} files contain hyphen characters:''') print('''\n'''.join(hyphen_files) + '''\n''') snake_case_ = [file for file in filepaths if os.sep not in file] if nodir_files: print(F'''{len(nodir_files)} files are not in a directory:''') print('''\n'''.join(nodir_files) + '''\n''') snake_case_ = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
216
0
class a__ : def __init__( self : Optional[Any],_A : int,_A : List[str]=None,_A : Tuple=None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = data SCREAMING_SNAKE_CASE_ : int = previous SCREAMING_SNAKE_CASE_ : Optional[int] = next_node def __str__( self : Tuple ): """simple docstring""" return F'{self.data}' def __UpperCamelCase ( self : str ): """simple docstring""" return self.data def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" return self.next def __UpperCamelCase ( self : Any ): """simple docstring""" return self.previous class a__ : def __init__( self : Optional[int],_A : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = head def __iter__( self : Any ): """simple docstring""" return self def __UpperCamelCase ( self : Dict ): """simple docstring""" if not self.current: raise StopIteration else: SCREAMING_SNAKE_CASE_ : List[Any] = self.current.get_data() SCREAMING_SNAKE_CASE_ : Optional[Any] = self.current.get_next() return value class a__ : def __init__( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = None # First node in list SCREAMING_SNAKE_CASE_ : str = None # Last node in list def __str__( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = self.head SCREAMING_SNAKE_CASE_ : Union[str, Any] = [] while current is not None: nodes.append(current.get_data() ) SCREAMING_SNAKE_CASE_ : Tuple = current.get_next() return " ".join(str(_A ) for node in nodes ) def __contains__( self : List[Any],_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.head while current: if current.get_data() == value: return True SCREAMING_SNAKE_CASE_ : List[str] = current.get_next() return False def __iter__( self : int ): """simple docstring""" return LinkedListIterator(self.head ) def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" if self.head: return self.head.get_data() return None def __UpperCamelCase ( self : Dict ): """simple docstring""" if self.tail: return self.tail.get_data() return None def __UpperCamelCase ( self : Tuple,_A : Node ): """simple docstring""" if self.head is None: SCREAMING_SNAKE_CASE_ : List[Any] = node SCREAMING_SNAKE_CASE_ : Union[str, Any] = node else: self.insert_before_node(self.head,_A ) def __UpperCamelCase ( self : Optional[Any],_A : Node ): """simple docstring""" if self.head is None: self.set_head(_A ) else: self.insert_after_node(self.tail,_A ) def __UpperCamelCase ( self : Optional[Any],_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = Node(_A ) if self.head is None: self.set_head(_A ) else: self.set_tail(_A ) def __UpperCamelCase ( self : Any,_A : Node,_A : Node ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = node SCREAMING_SNAKE_CASE_ : str = node.previous if node.get_previous() is None: SCREAMING_SNAKE_CASE_ : Optional[Any] = node_to_insert else: SCREAMING_SNAKE_CASE_ : Dict = node_to_insert SCREAMING_SNAKE_CASE_ : Any = node_to_insert def __UpperCamelCase ( self : List[Any],_A : Node,_A : Node ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = node SCREAMING_SNAKE_CASE_ : List[str] = node.next if node.get_next() is None: SCREAMING_SNAKE_CASE_ : Any = node_to_insert else: SCREAMING_SNAKE_CASE_ : str = node_to_insert SCREAMING_SNAKE_CASE_ : Any = node_to_insert def __UpperCamelCase ( self : str,_A : int,_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = 1 SCREAMING_SNAKE_CASE_ : Union[str, Any] = Node(_A ) SCREAMING_SNAKE_CASE_ : int = self.head while node: if current_position == position: self.insert_before_node(_A,_A ) return current_position += 1 SCREAMING_SNAKE_CASE_ : List[str] = node.next self.insert_after_node(self.tail,_A ) def __UpperCamelCase ( self : Union[str, Any],_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.head while node: if node.get_data() == item: return node SCREAMING_SNAKE_CASE_ : int = node.get_next() raise Exception("Node not found" ) def __UpperCamelCase ( self : List[str],_A : List[Any] ): """simple docstring""" if (node := self.get_node(_A )) is not None: if node == self.head: SCREAMING_SNAKE_CASE_ : Tuple = self.head.get_next() if node == self.tail: SCREAMING_SNAKE_CASE_ : int = self.tail.get_previous() self.remove_node_pointers(_A ) @staticmethod def __UpperCamelCase ( _A : Node ): """simple docstring""" if node.get_next(): SCREAMING_SNAKE_CASE_ : Union[str, Any] = node.previous if node.get_previous(): SCREAMING_SNAKE_CASE_ : List[Any] = node.next SCREAMING_SNAKE_CASE_ : str = None SCREAMING_SNAKE_CASE_ : str = None def __UpperCamelCase ( self : List[Any] ): """simple docstring""" return self.head is None def _snake_case ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
18
import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : Tuple = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} __lowerCamelCase : List[Any] = { '''vocab_file''': { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json''', '''allenai/longformer-large-4096''': ( '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json''' ), '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json''' ), }, '''merges_file''': { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt''', '''allenai/longformer-large-4096''': ( '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt''' ), '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt''' ), }, } __lowerCamelCase : int = { '''allenai/longformer-base-4096''': 40_96, '''allenai/longformer-large-4096''': 40_96, '''allenai/longformer-large-4096-finetuned-triviaqa''': 40_96, '''allenai/longformer-base-4096-extra.pos.embd.only''': 40_96, '''allenai/longformer-large-4096-extra.pos.embd.only''': 40_96, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def _snake_case ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) SCREAMING_SNAKE_CASE_ : str = bs[:] SCREAMING_SNAKE_CASE_ : Optional[int] = 0 for b in range(2**8 ): if b not in bs: bs.append(lowerCAmelCase ) cs.append(2**8 + n ) n += 1 SCREAMING_SNAKE_CASE_ : List[str] = [chr(lowerCAmelCase ) for n in cs] return dict(zip(lowerCAmelCase , lowerCAmelCase ) ) def _snake_case ( lowerCAmelCase : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = set() SCREAMING_SNAKE_CASE_ : Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char) ) SCREAMING_SNAKE_CASE_ : List[str] = char return pairs class a__ ( A__ ): A = VOCAB_FILES_NAMES A = PRETRAINED_VOCAB_FILES_MAP A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A = ['input_ids', 'attention_mask'] def __init__( self : Union[str, Any],_A : List[Any],_A : Tuple,_A : str="replace",_A : Optional[int]="<s>",_A : Dict="</s>",_A : Any="</s>",_A : Optional[Any]="<s>",_A : Union[str, Any]="<unk>",_A : int="<pad>",_A : Dict="<mask>",_A : int=False,**_A : Dict,): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else bos_token SCREAMING_SNAKE_CASE_ : Optional[int] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else eos_token SCREAMING_SNAKE_CASE_ : str = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else sep_token SCREAMING_SNAKE_CASE_ : Union[str, Any] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else cls_token SCREAMING_SNAKE_CASE_ : List[str] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else unk_token SCREAMING_SNAKE_CASE_ : Optional[Any] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else pad_token # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE_ : Dict = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else mask_token super().__init__( errors=_A,bos_token=_A,eos_token=_A,unk_token=_A,sep_token=_A,cls_token=_A,pad_token=_A,mask_token=_A,add_prefix_space=_A,**_A,) with open(_A,encoding="utf-8" ) as vocab_handle: SCREAMING_SNAKE_CASE_ : Tuple = json.load(_A ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = {v: k for k, v in self.encoder.items()} SCREAMING_SNAKE_CASE_ : Any = errors # how to handle errors in decoding SCREAMING_SNAKE_CASE_ : Optional[Any] = bytes_to_unicode() SCREAMING_SNAKE_CASE_ : str = {v: k for k, v in self.byte_encoder.items()} with open(_A,encoding="utf-8" ) as merges_handle: SCREAMING_SNAKE_CASE_ : int = merges_handle.read().split("\n" )[1:-1] SCREAMING_SNAKE_CASE_ : List[str] = [tuple(merge.split() ) for merge in bpe_merges] SCREAMING_SNAKE_CASE_ : Optional[int] = dict(zip(_A,range(len(_A ) ) ) ) SCREAMING_SNAKE_CASE_ : Any = {} SCREAMING_SNAKE_CASE_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions SCREAMING_SNAKE_CASE_ : List[Any] = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def __UpperCamelCase ( self : List[str] ): """simple docstring""" return len(self.encoder ) def __UpperCamelCase ( self : Tuple ): """simple docstring""" return dict(self.encoder,**self.added_tokens_encoder ) def __UpperCamelCase ( self : Any,_A : int ): """simple docstring""" if token in self.cache: return self.cache[token] SCREAMING_SNAKE_CASE_ : Union[str, Any] = tuple(_A ) SCREAMING_SNAKE_CASE_ : str = get_pairs(_A ) if not pairs: return token while True: SCREAMING_SNAKE_CASE_ : Tuple = min(_A,key=lambda _A : self.bpe_ranks.get(_A,float("inf" ) ) ) if bigram not in self.bpe_ranks: break SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Tuple = bigram SCREAMING_SNAKE_CASE_ : int = [] SCREAMING_SNAKE_CASE_ : Dict = 0 while i < len(_A ): try: SCREAMING_SNAKE_CASE_ : Tuple = word.index(_A,_A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) SCREAMING_SNAKE_CASE_ : str = j if word[i] == first and i < len(_A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 SCREAMING_SNAKE_CASE_ : Dict = tuple(_A ) SCREAMING_SNAKE_CASE_ : List[str] = new_word if len(_A ) == 1: break else: SCREAMING_SNAKE_CASE_ : Union[str, Any] = get_pairs(_A ) SCREAMING_SNAKE_CASE_ : List[str] = " ".join(_A ) SCREAMING_SNAKE_CASE_ : Any = word return word def __UpperCamelCase ( self : Dict,_A : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = [] for token in re.findall(self.pat,_A ): SCREAMING_SNAKE_CASE_ : Any = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_A ).split(" " ) ) return bpe_tokens def __UpperCamelCase ( self : Optional[int],_A : str ): """simple docstring""" return self.encoder.get(_A,self.encoder.get(self.unk_token ) ) def __UpperCamelCase ( self : Tuple,_A : str ): """simple docstring""" return self.decoder.get(_A ) def __UpperCamelCase ( self : List[str],_A : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = "".join(_A ) SCREAMING_SNAKE_CASE_ : Tuple = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8",errors=self.errors ) return text def __UpperCamelCase ( self : List[Any],_A : str,_A : Optional[str] = None ): """simple docstring""" if not os.path.isdir(_A ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return SCREAMING_SNAKE_CASE_ : Tuple = os.path.join( _A,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) SCREAMING_SNAKE_CASE_ : Any = os.path.join( _A,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_A,"w",encoding="utf-8" ) as f: f.write(json.dumps(self.encoder,indent=2,sort_keys=_A,ensure_ascii=_A ) + "\n" ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = 0 with open(_A,"w",encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items(),key=lambda _A : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' " Please check that the tokenizer is not corrupted!" ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = token_index writer.write(" ".join(_A ) + "\n" ) index += 1 return vocab_file, merge_file def __UpperCamelCase ( self : Optional[Any],_A : List[int],_A : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE_ : str = [self.cls_token_id] SCREAMING_SNAKE_CASE_ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __UpperCamelCase ( self : str,_A : List[int],_A : Optional[List[int]] = None,_A : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A,token_ids_a=_A,already_has_special_tokens=_A ) if token_ids_a is None: return [1] + ([0] * len(_A )) + [1] return [1] + ([0] * len(_A )) + [1, 1] + ([0] * len(_A )) + [1] def __UpperCamelCase ( self : Any,_A : List[int],_A : Optional[List[int]] = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __UpperCamelCase ( self : Any,_A : Union[str, Any],_A : Any=False,**_A : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = kwargs.pop("add_prefix_space",self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_A ) > 0 and not text[0].isspace()): SCREAMING_SNAKE_CASE_ : str = " " + text return (text, kwargs)
18
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : str = logging.get_logger(__name__) def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase=False ): __UpperCAmelCase : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((F"blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __UpperCAmelCase : Union[str, Any] = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase=False ): for i in range(config.num_hidden_layers ): if base_model: __UpperCAmelCase : Optional[int] = "" else: __UpperCAmelCase : List[Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __UpperCAmelCase : Optional[int] = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) __UpperCAmelCase : List[Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __UpperCAmelCase : int = in_proj_weight[ : config.hidden_size, : ] __UpperCAmelCase : str = in_proj_bias[: config.hidden_size] __UpperCAmelCase : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __UpperCAmelCase : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] __UpperCAmelCase : Tuple = in_proj_bias[-config.hidden_size :] def __UpperCamelCase ( _UpperCAmelCase ): __UpperCAmelCase : Tuple = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_UpperCAmelCase, _UpperCAmelCase ) def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = dct.pop(_UpperCAmelCase ) __UpperCAmelCase : Dict = val def __UpperCamelCase ( ): __UpperCAmelCase : Any = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCAmelCase : int = Image.open(requests.get(_UpperCAmelCase, stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : int = ViTConfig() __UpperCAmelCase : Tuple = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : Union[str, Any] = int(vit_name[-12:-10] ) __UpperCAmelCase : int = int(vit_name[-9:-6] ) else: __UpperCAmelCase : str = 1000 __UpperCAmelCase : List[str] = "huggingface/label-files" __UpperCAmelCase : List[str] = "imagenet-1k-id2label.json" __UpperCAmelCase : Optional[int] = json.load(open(hf_hub_download(_UpperCAmelCase, _UpperCAmelCase, repo_type="dataset" ), "r" ) ) __UpperCAmelCase : Optional[Any] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} __UpperCAmelCase : Union[str, Any] = idalabel __UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} __UpperCAmelCase : List[str] = int(vit_name[-6:-4] ) __UpperCAmelCase : Dict = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith("tiny" ): __UpperCAmelCase : Optional[Any] = 192 __UpperCAmelCase : int = 768 __UpperCAmelCase : List[Any] = 12 __UpperCAmelCase : str = 3 elif vit_name[9:].startswith("small" ): __UpperCAmelCase : Dict = 384 __UpperCAmelCase : Optional[int] = 1536 __UpperCAmelCase : int = 12 __UpperCAmelCase : str = 6 else: pass else: if vit_name[4:].startswith("small" ): __UpperCAmelCase : Any = 768 __UpperCAmelCase : List[Any] = 2304 __UpperCAmelCase : str = 8 __UpperCAmelCase : Tuple = 8 elif vit_name[4:].startswith("base" ): pass elif vit_name[4:].startswith("large" ): __UpperCAmelCase : Dict = 1024 __UpperCAmelCase : Any = 4096 __UpperCAmelCase : Optional[Any] = 24 __UpperCAmelCase : Optional[int] = 16 elif vit_name[4:].startswith("huge" ): __UpperCAmelCase : Optional[Any] = 1280 __UpperCAmelCase : Any = 5120 __UpperCAmelCase : Optional[Any] = 32 __UpperCAmelCase : str = 16 # load original model from timm __UpperCAmelCase : Optional[int] = timm.create_model(_UpperCAmelCase, pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys __UpperCAmelCase : str = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) __UpperCAmelCase : Any = create_rename_keys(_UpperCAmelCase, _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": __UpperCAmelCase : Optional[int] = ViTModel(_UpperCAmelCase ).eval() else: __UpperCAmelCase : Tuple = ViTForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: __UpperCAmelCase : Tuple = DeiTImageProcessor(size=config.image_size ) else: __UpperCAmelCase : Optional[int] = ViTImageProcessor(size=config.image_size ) __UpperCAmelCase : List[str] = image_processor(images=prepare_img(), return_tensors="pt" ) __UpperCAmelCase : Any = encoding["pixel_values"] __UpperCAmelCase : Dict = model(_UpperCAmelCase ) if base_model: __UpperCAmelCase : Tuple = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase, outputs.pooler_output, atol=1E-3 ) else: __UpperCAmelCase : Any = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase, outputs.logits, atol=1E-3 ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(F"Saving model {vit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_UpperCAmelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_patch16_224", type=str, help="Name of the ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) lowerCAmelCase__ : str = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
354
'''simple docstring''' from __future__ import annotations import collections import pprint from pathlib import Path def __UpperCamelCase ( _UpperCAmelCase ): return "".join(sorted(_UpperCAmelCase ) ) def __UpperCamelCase ( _UpperCAmelCase ): return word_by_signature[signature(_UpperCAmelCase )] lowerCAmelCase__ : str = Path(__file__).parent.joinpath("words.txt").read_text(encoding="utf-8") lowerCAmelCase__ : Optional[int] = sorted({word.strip().lower() for word in data.splitlines()}) lowerCAmelCase__ : Tuple = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": lowerCAmelCase__ : int = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("anagrams.txt", "w") as file: file.write("all_anagrams = \n ") file.write(pprint.pformat(all_anagrams))
37
0
from timeit import timeit def __snake_case ( _UpperCAmelCase ): if number < 0: raise ValueError('''the value of input must not be negative''' ) __a = 0 while number: number &= number - 1 result += 1 return result def __snake_case ( _UpperCAmelCase ): if number < 0: raise ValueError('''the value of input must not be negative''' ) __a = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def __snake_case ( ): def do_benchmark(_UpperCAmelCase ) -> None: __a = '''import __main__ as z''' print(f'Benchmark when {number = }:' ) print(f'{get_set_bits_count_using_modulo_operator(_UpperCAmelCase ) = }' ) __a = timeit('''z.get_set_bits_count_using_modulo_operator(25)''' , setup=_UpperCAmelCase ) print(f'timeit() runs in {timing} seconds' ) print(f'{get_set_bits_count_using_brian_kernighans_algorithm(_UpperCAmelCase ) = }' ) __a = timeit( '''z.get_set_bits_count_using_brian_kernighans_algorithm(25)''' , setup=_UpperCAmelCase , ) print(f'timeit() runs in {timing} seconds' ) for number in (25, 37, 58, 0): do_benchmark(_UpperCAmelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
49
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class _A ( __UpperCAmelCase ): UpperCamelCase__ : Tuple = (DDPMParallelScheduler,) def _lowerCamelCase ( self : int , **__SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' __a = { '''num_train_timesteps''': 1_000, '''beta_start''': 0.00_01, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''variance_type''': '''fixed_small''', '''clip_sample''': True, } config.update(**__SCREAMING_SNAKE_CASE) return config def _lowerCamelCase ( self : List[str]): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Tuple): '''simple docstring''' for beta_start, beta_end in zip([0.00_01, 0.0_01, 0.01, 0.1] , [0.0_02, 0.02, 0.2, 2]): self.check_over_configs(beta_start=__SCREAMING_SNAKE_CASE , beta_end=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[str]): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Any): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[str]): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : str): '''simple docstring''' self.check_over_configs(thresholding=__SCREAMING_SNAKE_CASE) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=__SCREAMING_SNAKE_CASE , prediction_type=__SCREAMING_SNAKE_CASE , sample_max_value=__SCREAMING_SNAKE_CASE , ) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[str]): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : str): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) assert torch.sum(torch.abs(scheduler._get_variance(0) - 0.0)) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487) - 0.0_09_79)) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999) - 0.02)) < 1E-5 def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = len(__SCREAMING_SNAKE_CASE) __a = self.dummy_model() __a = self.dummy_sample_deter __a = self.dummy_sample_deter + 0.1 __a = self.dummy_sample_deter - 0.1 __a = samplea.shape[0] __a = torch.stack([samplea, samplea, samplea] , dim=0) __a = torch.arange(__SCREAMING_SNAKE_CASE)[0:3, None].repeat(1 , __SCREAMING_SNAKE_CASE) __a = model(samples.flatten(0 , 1) , timesteps.flatten(0 , 1)) __a = scheduler.batch_step_no_noise(__SCREAMING_SNAKE_CASE , timesteps.flatten(0 , 1) , samples.flatten(0 , 1)) __a = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE)) __a = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE)) assert abs(result_sum.item() - 11_53.18_33) < 1E-2 assert abs(result_mean.item() - 0.50_05) < 1E-3 def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = len(__SCREAMING_SNAKE_CASE) __a = self.dummy_model() __a = self.dummy_sample_deter __a = torch.manual_seed(0) for t in reversed(range(__SCREAMING_SNAKE_CASE)): # 1. predict noise residual __a = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # 2. predict previous mean of sample x_t-1 __a = scheduler.step(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE).prev_sample __a = pred_prev_sample __a = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE)) __a = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE)) assert abs(result_sum.item() - 2_58.96_06) < 1E-2 assert abs(result_mean.item() - 0.33_72) < 1E-3 def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config(prediction_type='''v_prediction''') __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = len(__SCREAMING_SNAKE_CASE) __a = self.dummy_model() __a = self.dummy_sample_deter __a = torch.manual_seed(0) for t in reversed(range(__SCREAMING_SNAKE_CASE)): # 1. predict noise residual __a = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # 2. predict previous mean of sample x_t-1 __a = scheduler.step(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE).prev_sample __a = pred_prev_sample __a = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE)) __a = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE)) assert abs(result_sum.item() - 2_02.02_96) < 1E-2 assert abs(result_mean.item() - 0.26_31) < 1E-3 def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE) __a = scheduler.timesteps for i, timestep in enumerate(__SCREAMING_SNAKE_CASE): if i == len(__SCREAMING_SNAKE_CASE) - 1: __a = -1 else: __a = timesteps[i + 1] __a = scheduler.previous_timestep(__SCREAMING_SNAKE_CASE) __a = prev_t.item() self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = [100, 87, 50, 51, 0] with self.assertRaises(__SCREAMING_SNAKE_CASE , msg='''`custom_timesteps` must be in descending order.'''): scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = [100, 87, 50, 1, 0] __a = len(__SCREAMING_SNAKE_CASE) with self.assertRaises(__SCREAMING_SNAKE_CASE , msg='''Can only pass one of `num_inference_steps` or `custom_timesteps`.'''): scheduler.set_timesteps(num_inference_steps=__SCREAMING_SNAKE_CASE , timesteps=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = self.scheduler_classes[0] __a = self.get_scheduler_config() __a = scheduler_class(**__SCREAMING_SNAKE_CASE) __a = [scheduler.config.num_train_timesteps] with self.assertRaises( __SCREAMING_SNAKE_CASE , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE)
49
1
"""simple docstring""" import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class lowercase__ ( snake_case__ ): def __init__( self : int , snake_case__ : Union[str, "sqlalchemy.sql.Selectable"] , snake_case__ : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] , snake_case__ : Optional[Features] = None , snake_case__ : str = None , snake_case__ : bool = False , **snake_case__ : Any , ): super().__init__(features=snake_case__ , cache_dir=snake_case__ , keep_in_memory=snake_case__ , **snake_case__ ) lowerCamelCase_ : int =Sql( cache_dir=snake_case__ , features=snake_case__ , sql=snake_case__ , con=snake_case__ , **snake_case__ , ) def UpperCAmelCase__ ( self : Optional[int] ): lowerCamelCase_ : str =None lowerCamelCase_ : Any =None lowerCamelCase_ : Tuple =None lowerCamelCase_ : str =None self.builder.download_and_prepare( download_config=snake_case__ , download_mode=snake_case__ , verification_mode=snake_case__ , base_path=snake_case__ , ) # Build dataset for splits lowerCamelCase_ : List[Any] =self.builder.as_dataset( split="train" , verification_mode=snake_case__ , in_memory=self.keep_in_memory ) return dataset class lowercase__ : def __init__( self : Dict , snake_case__ : Dataset , snake_case__ : str , snake_case__ : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] , snake_case__ : Optional[int] = None , snake_case__ : Optional[int] = None , **snake_case__ : Union[str, Any] , ): if num_proc is not None and num_proc <= 0: raise ValueError(F"""num_proc {num_proc} must be an integer > 0.""" ) lowerCamelCase_ : Tuple =dataset lowerCamelCase_ : Dict =name lowerCamelCase_ : Optional[Any] =con lowerCamelCase_ : List[Any] =batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE lowerCamelCase_ : List[str] =num_proc lowerCamelCase_ : Tuple =to_sql_kwargs def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : List[str] =self.to_sql_kwargs.pop("sql" , snake_case__ ) lowerCamelCase_ : Optional[int] =self.to_sql_kwargs.pop("con" , snake_case__ ) lowerCamelCase_ : int =self.to_sql_kwargs.pop("index" , snake_case__ ) lowerCamelCase_ : List[str] =self._write(index=snake_case__ , **self.to_sql_kwargs ) return written def UpperCAmelCase__ ( self : List[Any] , snake_case__ : Dict ): lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Tuple =args lowerCamelCase_ : Union[str, Any] ={**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs lowerCamelCase_ : int =query_table( table=self.dataset.data , key=slice(snake_case__ , offset + self.batch_size ) , indices=self.dataset._indices , ) lowerCamelCase_ : str =batch.to_pandas() lowerCamelCase_ : List[Any] =df.to_sql(self.name , self.con , index=snake_case__ , **snake_case__ ) return num_rows or len(snake_case__ ) def UpperCAmelCase__ ( self : List[str] , snake_case__ : Optional[Any] , **snake_case__ : Optional[Any] ): lowerCamelCase_ : int =0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: lowerCamelCase_ , lowerCamelCase_ : Optional[int] =len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql , [(offset, index, to_sql_kwargs) for offset in range(0 , snake_case__ , snake_case__ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += num_rows return written
209
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[Any] = { 'configuration_lilt': ['LILT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LiltConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : List[str] = [ 'LILT_PRETRAINED_MODEL_ARCHIVE_LIST', 'LiltForQuestionAnswering', 'LiltForSequenceClassification', 'LiltForTokenClassification', 'LiltModel', 'LiltPreTrainedModel', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys A__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
209
1
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): UpperCAmelCase_ :int = CustomTokenizer pass
84
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class snake_case__(_UpperCamelCase , unittest.TestCase ): """simple docstring""" lowercase_ = LEDTokenizer lowercase_ = LEDTokenizerFast lowercase_ = True def snake_case ( self : str ): super().setUp() lowercase__ : Tuple = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowercase__ : Any = dict(zip(SCREAMING_SNAKE_CASE , range(len(SCREAMING_SNAKE_CASE ) ) ) ) lowercase__ : str = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowercase__ : List[str] = {"unk_token": "<unk>"} lowercase__ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(SCREAMING_SNAKE_CASE ) ) def snake_case ( self : Union[str, Any] , **SCREAMING_SNAKE_CASE : int ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE ) def snake_case ( self : Optional[int] , **SCREAMING_SNAKE_CASE : Tuple ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE ) def snake_case ( self : str , SCREAMING_SNAKE_CASE : Optional[int] ): return "lower newer", "lower newer" @cached_property def snake_case ( self : List[str] ): return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def snake_case ( self : Union[str, Any] ): return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def snake_case ( self : str ): lowercase__ : List[str] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowercase__ : List[str] = [0, 250, 251, 17_818, 13, 39_186, 1_938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : str = tokenizer(SCREAMING_SNAKE_CASE , max_length=len(SCREAMING_SNAKE_CASE ) , padding=SCREAMING_SNAKE_CASE , return_tensors="pt" ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) lowercase__ : Optional[Any] = batch.input_ids.tolist()[0] self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) @require_torch def snake_case ( self : Optional[int] ): lowercase__ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : Any = tokenizer(SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE , return_tensors="pt" ) self.assertIn("input_ids" , SCREAMING_SNAKE_CASE ) self.assertIn("attention_mask" , SCREAMING_SNAKE_CASE ) self.assertNotIn("labels" , SCREAMING_SNAKE_CASE ) self.assertNotIn("decoder_attention_mask" , SCREAMING_SNAKE_CASE ) @require_torch def snake_case ( self : List[str] ): lowercase__ : List[str] = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : Any = tokenizer(text_target=SCREAMING_SNAKE_CASE , max_length=32 , padding="max_length" , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) @require_torch def snake_case ( self : Optional[int] ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : Tuple = tokenizer( ["I am a small frog" * 1_024, "I am a small frog"] , padding=SCREAMING_SNAKE_CASE , truncation=SCREAMING_SNAKE_CASE , return_tensors="pt" ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) self.assertEqual(batch.input_ids.shape , (2, 5_122) ) @require_torch def snake_case ( self : Tuple ): lowercase__ : int = ["A long paragraph for summarization."] lowercase__ : Union[str, Any] = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : Tuple = tokenizer(SCREAMING_SNAKE_CASE , return_tensors="pt" ) lowercase__ : List[str] = tokenizer(text_target=SCREAMING_SNAKE_CASE , return_tensors="pt" ) lowercase__ : Union[str, Any] = inputs["input_ids"] lowercase__ : List[str] = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def snake_case ( self : Dict ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowercase__ : Optional[int] = ["Summary of the text.", "Another summary."] lowercase__ : List[Any] = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowercase__ : Union[str, Any] = tokenizer(SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE ) lowercase__ : List[str] = [[0] * len(SCREAMING_SNAKE_CASE ) for x in encoded_output["input_ids"]] lowercase__ : Optional[int] = tokenizer.pad(SCREAMING_SNAKE_CASE ) self.assertSequenceEqual(outputs["global_attention_mask"] , SCREAMING_SNAKE_CASE ) def snake_case ( self : List[str] ): pass def snake_case ( self : str ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowercase__ : Tuple = self.rust_tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) lowercase__ : List[str] = self.tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) lowercase__ : Optional[Any] = "A, <mask> AllenNLP sentence." lowercase__ : Any = tokenizer_r.encode_plus(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE , return_token_type_ids=SCREAMING_SNAKE_CASE ) lowercase__ : Optional[Any] = tokenizer_p.encode_plus(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE , return_token_type_ids=SCREAMING_SNAKE_CASE ) self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) lowercase__ : Optional[Any] = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowercase__ : str = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( SCREAMING_SNAKE_CASE , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( SCREAMING_SNAKE_CASE , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
130
0
def __magic_name__ ( __lowerCAmelCase : Dict , __lowerCAmelCase : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[int] ) -> List[Any]: if index == r: for j in range(__lowerCAmelCase ): print(data[j] , end=''' ''' ) print(''' ''' ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location __lowerCamelCase = arr[i] combination_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , index + 1 , __lowerCAmelCase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def __magic_name__ ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[Any] ) -> Optional[Any]: # A temporary array to store all combination one by one __lowerCamelCase = [0] * r # Print all combination using temporary array 'data[]' combination_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , 0 , __lowerCAmelCase , 0 ) if __name__ == "__main__": # Driver code to check the function above SCREAMING_SNAKE_CASE__ : int = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
339
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__ : Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Optional[Any] = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } SCREAMING_SNAKE_CASE__ : str = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } SCREAMING_SNAKE_CASE__ : int = {"facebook/blenderbot_small-90M": 512} def __magic_name__ ( __lowerCAmelCase : List[str] ) -> Tuple: __lowerCamelCase = set() __lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCamelCase = char __lowerCamelCase = set(__lowerCAmelCase ) return pairs class lowerCAmelCase__ ( __lowercase ): a__ : List[Any] = VOCAB_FILES_NAMES a__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP a__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Dict = ["""input_ids""", """attention_mask"""] def __init__( self : str , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple="__start__" , SCREAMING_SNAKE_CASE__ : Tuple="__end__" , SCREAMING_SNAKE_CASE__ : List[str]="__unk__" , SCREAMING_SNAKE_CASE__ : str="__null__" , **SCREAMING_SNAKE_CASE__ : Optional[Any] , ) -> Optional[Any]: super().__init__(unk_token=SCREAMING_SNAKE_CASE__ , bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , pad_token=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) with open(SCREAMING_SNAKE_CASE__ , encoding='''utf-8''' ) as vocab_handle: __lowerCamelCase = json.load(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = {v: k for k, v in self.encoder.items()} with open(SCREAMING_SNAKE_CASE__ , encoding='''utf-8''' ) as merges_handle: __lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] __lowerCamelCase = [tuple(merge.split() ) for merge in merges] __lowerCamelCase = dict(zip(SCREAMING_SNAKE_CASE__ , range(len(SCREAMING_SNAKE_CASE__ ) ) ) ) __lowerCamelCase = {} @property def __A ( self : Dict ) -> int: return len(self.encoder ) def __A ( self : str ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def __A ( self : Any , SCREAMING_SNAKE_CASE__ : str ) -> str: if token in self.cache: return self.cache[token] __lowerCamelCase = re.sub('''([.,!?()])''' , R''' \1''' , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = re.sub('''(\')''' , R''' \1 ''' , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = re.sub(R'''\s{2,}''' , ''' ''' , SCREAMING_SNAKE_CASE__ ) if "\n" in token: __lowerCamelCase = token.replace('''\n''' , ''' __newln__''' ) __lowerCamelCase = token.split(''' ''' ) __lowerCamelCase = [] for token in tokens: if not len(SCREAMING_SNAKE_CASE__ ): continue __lowerCamelCase = token.lower() __lowerCamelCase = tuple(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) __lowerCamelCase = get_pairs(SCREAMING_SNAKE_CASE__ ) if not pairs: words.append(SCREAMING_SNAKE_CASE__ ) continue while True: __lowerCamelCase = min(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : self.bpe_ranks.get(SCREAMING_SNAKE_CASE__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break __lowerCamelCase , __lowerCamelCase = bigram __lowerCamelCase = [] __lowerCamelCase = 0 while i < len(SCREAMING_SNAKE_CASE__ ): try: __lowerCamelCase = word.index(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) new_word.extend(word[i:j] ) __lowerCamelCase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(SCREAMING_SNAKE_CASE__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowerCamelCase = tuple(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = new_word if len(SCREAMING_SNAKE_CASE__ ) == 1: break else: __lowerCamelCase = get_pairs(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = '''@@ '''.join(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = word[:-4] __lowerCamelCase = word words.append(SCREAMING_SNAKE_CASE__ ) return " ".join(SCREAMING_SNAKE_CASE__ ) def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: __lowerCamelCase = [] __lowerCamelCase = re.findall(R'''\S+\n?''' , SCREAMING_SNAKE_CASE__ ) for token in words: split_tokens.extend(list(self.bpe(SCREAMING_SNAKE_CASE__ ).split(''' ''' ) ) ) return split_tokens def __A ( self : str , SCREAMING_SNAKE_CASE__ : str ) -> int: __lowerCamelCase = token.lower() return self.encoder.get(SCREAMING_SNAKE_CASE__ , self.encoder.get(self.unk_token ) ) def __A ( self : Dict , SCREAMING_SNAKE_CASE__ : int ) -> str: return self.decoder.get(SCREAMING_SNAKE_CASE__ , self.unk_token ) def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str] ) -> str: __lowerCamelCase = ''' '''.join(SCREAMING_SNAKE_CASE__ ).replace('''@@ ''' , '''''' ).strip() return out_string def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(SCREAMING_SNAKE_CASE__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __lowerCamelCase = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(SCREAMING_SNAKE_CASE__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=SCREAMING_SNAKE_CASE__ , ensure_ascii=SCREAMING_SNAKE_CASE__ ) + '''\n''' ) __lowerCamelCase = 0 with open(SCREAMING_SNAKE_CASE__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda SCREAMING_SNAKE_CASE__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' ''' Please check that the tokenizer is not corrupted!''' ) __lowerCamelCase = token_index writer.write(''' '''.join(SCREAMING_SNAKE_CASE__ ) + '''\n''' ) index += 1 return vocab_file, merge_file
339
1
'''simple docstring''' import argparse import os import re import packaging.version UpperCAmelCase_ = 'examples/' UpperCAmelCase_ = { 'examples': (re.compile(r'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(r'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(r'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), r'\1version="VERSION",'), 'doc': (re.compile(r'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } UpperCAmelCase_ = { 'init': 'src/diffusers/__init__.py', 'setup': 'setup.py', } UpperCAmelCase_ = 'README.md' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase__ = f.read() UpperCAmelCase__ , UpperCAmelCase__ = REPLACE_PATTERNS[pattern] UpperCAmelCase__ = replace.replace("""VERSION""" , SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = re_pattern.sub(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) with open(SCREAMING_SNAKE_CASE__ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' for folder, directories, fnames in os.walk(SCREAMING_SNAKE_CASE__ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ , pattern="""examples""" ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : str=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if not patch: update_version_in_examples(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = """🤗 Transformers currently provides the following architectures""" UpperCAmelCase__ = """1. Want to contribute a new model?""" with open(SCREAMING_SNAKE_CASE__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase__ = f.readlines() # Find the start of the list. UpperCAmelCase__ = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase__ = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): UpperCAmelCase__ = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(SCREAMING_SNAKE_CASE__ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: UpperCAmelCase__ = f.read() UpperCAmelCase__ = REPLACE_PATTERNS["""init"""][0].search(SCREAMING_SNAKE_CASE__ ).groups()[0] return packaging.version.parse(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[Any]=False ): '''simple docstring''' UpperCAmelCase__ = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: UpperCAmelCase__ = default_version.base_version elif patch: UpperCAmelCase__ = F'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: UpperCAmelCase__ = F'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. UpperCAmelCase__ = input(F'''Which version are you releasing? [{default_version}]''' ) if len(SCREAMING_SNAKE_CASE__ ) == 0: UpperCAmelCase__ = default_version print(F'''Updating version to {version}.''' ) global_version_update(SCREAMING_SNAKE_CASE__ , patch=SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = get_version() UpperCAmelCase__ = F'''{current_version.major}.{current_version.minor + 1}.0.dev0''' UpperCAmelCase__ = current_version.base_version # Check with the user we got that right. UpperCAmelCase__ = input(F'''Which version are we developing now? [{dev_version}]''' ) if len(SCREAMING_SNAKE_CASE__ ) == 0: UpperCAmelCase__ = dev_version print(F'''Updating version to {version}.''' ) global_version_update(SCREAMING_SNAKE_CASE__ ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') UpperCAmelCase_ = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
346
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' @register_to_config def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : float , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : str , _UpperCAmelCase : bool = False , ): """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = False UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) UpperCAmelCase__ = TaConfig( vocab_size=_UpperCAmelCase , d_model=_UpperCAmelCase , num_heads=_UpperCAmelCase , d_kv=_UpperCAmelCase , d_ff=_UpperCAmelCase , dropout_rate=_UpperCAmelCase , feed_forward_proj=_UpperCAmelCase , is_decoder=_UpperCAmelCase , is_encoder_decoder=_UpperCAmelCase , ) UpperCAmelCase__ = nn.ModuleList() for lyr_num in range(_UpperCAmelCase ): UpperCAmelCase__ = TaBlock(_UpperCAmelCase ) self.encoders.append(_UpperCAmelCase ) UpperCAmelCase__ = TaLayerNorm(_UpperCAmelCase ) UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Dict , _UpperCAmelCase : int , _UpperCAmelCase : str ): """simple docstring""" UpperCAmelCase__ = self.token_embedder(_UpperCAmelCase ) UpperCAmelCase__ = encoder_input_tokens.shape[1] UpperCAmelCase__ = torch.arange(_UpperCAmelCase , device=encoder_input_tokens.device ) x += self.position_encoding(_UpperCAmelCase ) UpperCAmelCase__ = self.dropout_pre(_UpperCAmelCase ) # inverted the attention mask UpperCAmelCase__ = encoder_input_tokens.size() UpperCAmelCase__ = self.get_extended_attention_mask(_UpperCAmelCase , _UpperCAmelCase ) for lyr in self.encoders: UpperCAmelCase__ = lyr(_UpperCAmelCase , _UpperCAmelCase )[0] UpperCAmelCase__ = self.layer_norm(_UpperCAmelCase ) return self.dropout_post(_UpperCAmelCase ), encoder_inputs_mask
346
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case : Union[str, Any] = { "configuration_swinv2": ["SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Swinv2Config"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : str = [ "SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST", "Swinv2ForImageClassification", "Swinv2ForMaskedImageModeling", "Swinv2Model", "Swinv2PreTrainedModel", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys snake_case : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
41
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) snake_case : List[str] = {"configuration_beit": ["BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BeitConfig", "BeitOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : int = ["BeitFeatureExtractor"] snake_case : Optional[int] = ["BeitImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : List[Any] = [ "BEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "BeitForImageClassification", "BeitForMaskedImageModeling", "BeitForSemanticSegmentation", "BeitModel", "BeitPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : Dict = [ "FlaxBeitForImageClassification", "FlaxBeitForMaskedImageModeling", "FlaxBeitModel", "FlaxBeitPreTrainedModel", ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys snake_case : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
41
1
from typing import Any def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): _validation( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ) # Creates data structures and fill initial step __SCREAMING_SNAKE_CASE : dict = {} __SCREAMING_SNAKE_CASE : dict = {} for state in states_space: __SCREAMING_SNAKE_CASE : List[Any] = observations_space[0] __SCREAMING_SNAKE_CASE : str = ( initial_probabilities[state] * emission_probabilities[state][observation] ) __SCREAMING_SNAKE_CASE : List[Any] = None # Fills the data structure with the probabilities of # different transitions and pointers to previous states for o in range(1 , len(lowercase__ ) ): __SCREAMING_SNAKE_CASE : List[Any] = observations_space[o] __SCREAMING_SNAKE_CASE : Optional[Any] = observations_space[o - 1] for state in states_space: # Calculates the argmax for probability function __SCREAMING_SNAKE_CASE : int = '''''' __SCREAMING_SNAKE_CASE : Any = -1 for k_state in states_space: __SCREAMING_SNAKE_CASE : int = ( probabilities[(k_state, prior_observation)] * transition_probabilities[k_state][state] * emission_probabilities[state][observation] ) if probability > max_probability: __SCREAMING_SNAKE_CASE : Tuple = probability __SCREAMING_SNAKE_CASE : Union[str, Any] = k_state # Update probabilities and pointers dicts __SCREAMING_SNAKE_CASE : str = ( probabilities[(arg_max, prior_observation)] * transition_probabilities[arg_max][state] * emission_probabilities[state][observation] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = arg_max # The final observation __SCREAMING_SNAKE_CASE : Union[str, Any] = observations_space[len(lowercase__ ) - 1] # argmax for given final observation __SCREAMING_SNAKE_CASE : Tuple = '''''' __SCREAMING_SNAKE_CASE : Optional[int] = -1 for k_state in states_space: __SCREAMING_SNAKE_CASE : Dict = probabilities[(k_state, final_observation)] if probability > max_probability: __SCREAMING_SNAKE_CASE : int = probability __SCREAMING_SNAKE_CASE : Optional[int] = k_state __SCREAMING_SNAKE_CASE : Optional[int] = arg_max # Process pointers backwards __SCREAMING_SNAKE_CASE : List[Any] = last_state __SCREAMING_SNAKE_CASE : Optional[Any] = [] for o in range(len(lowercase__ ) - 1 , -1 , -1 ): result.append(lowercase__ ) __SCREAMING_SNAKE_CASE : Tuple = pointers[previous, observations_space[o]] result.reverse() return result def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): _validate_not_empty( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ) _validate_lists(lowercase__ , lowercase__ ) _validate_dicts( lowercase__ , lowercase__ , lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): if not all( [ observations_space, states_space, initial_probabilities, transition_probabilities, emission_probabilities, ] ): raise ValueError('''There\'s an empty parameter''' ) def _UpperCamelCase ( lowercase__ , lowercase__ ): _validate_list(lowercase__ , '''observations_space''' ) _validate_list(lowercase__ , '''states_space''' ) def _UpperCamelCase ( lowercase__ , lowercase__ ): if not isinstance(_object , lowercase__ ): __SCREAMING_SNAKE_CASE : str = F'''{var_name} must be a list''' raise ValueError(lowercase__ ) else: for x in _object: if not isinstance(lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : str = F'''{var_name} must be a list of strings''' raise ValueError(lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , ): _validate_dict(lowercase__ , '''initial_probabilities''' , lowercase__ ) _validate_nested_dict(lowercase__ , '''transition_probabilities''' ) _validate_nested_dict(lowercase__ , '''emission_probabilities''' ) def _UpperCamelCase ( lowercase__ , lowercase__ ): _validate_dict(_object , lowercase__ , lowercase__ ) for x in _object.values(): _validate_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): if not isinstance(_object , lowercase__ ): __SCREAMING_SNAKE_CASE : str = F'''{var_name} must be a dict''' raise ValueError(lowercase__ ) if not all(isinstance(lowercase__ , lowercase__ ) for x in _object ): __SCREAMING_SNAKE_CASE : Dict = F'''{var_name} all keys must be strings''' raise ValueError(lowercase__ ) if not all(isinstance(lowercase__ , lowercase__ ) for x in _object.values() ): __SCREAMING_SNAKE_CASE : Tuple = '''nested dictionary ''' if nested else '''''' __SCREAMING_SNAKE_CASE : Optional[Any] = F'''{var_name} {nested_text}all values must be {value_type.__name__}''' raise ValueError(lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
9
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _lowerCAmelCase = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' _lowerCAmelCase = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' _lowerCAmelCase = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return float((preds == labels).mean() ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase="binary" ): """simple docstring""" lowerCAmelCase__ : Any = simple_accuracy(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : Tuple = float(fa_score(y_true=UpperCamelCase , y_pred=UpperCamelCase , average=UpperCamelCase ) ) return { "accuracy": acc, "f1": fa, } def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : List[str] = {} for id_pred, label in zip(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : str = f"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" lowerCAmelCase__ : Dict = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase__ : Optional[int] = [(pred, label)] lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = zip(*UpperCamelCase ) lowerCAmelCase__ : List[Any] = fa_score(y_true=UpperCamelCase , y_pred=UpperCamelCase , average="""macro""" ) fas.append(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = int(sum(pred == label for pred, label in preds_labels ) == len(UpperCamelCase ) ) ems.append(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = float(sum(UpperCamelCase ) / len(UpperCamelCase ) ) lowerCAmelCase__ : List[Any] = sum(UpperCamelCase ) / len(UpperCamelCase ) lowerCAmelCase__ : Dict = float(fa_score(y_true=UpperCamelCase , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_( datasets.Metric ): '''simple docstring''' def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features(self._get_feature_types() ) ,codebase_urls=[] ,reference_urls=[] ,format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None ,) def UpperCAmelCase_ ( self ) -> str: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> Any: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__UpperCAmelCase ,__UpperCAmelCase )} elif self.config_name == "cb": return acc_and_fa(__UpperCAmelCase ,__UpperCAmelCase ,fa_avg="""macro""" ) elif self.config_name == "record": lowerCAmelCase__ : Optional[Any] = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] lowerCAmelCase__ : Union[str, Any] = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__UpperCAmelCase ,__UpperCAmelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__UpperCAmelCase ,__UpperCAmelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__UpperCAmelCase ,__UpperCAmelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
37
0
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __snake_case = """src/diffusers""" __snake_case = """.""" # This is to make sure the diffusers module imported is the one in the repo. __snake_case = importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) __snake_case = spec.loader.load_module() def __lowerCAmelCase ( lowercase : str , lowercase : int ) -> Dict: """simple docstring""" return line.startswith(lowercase ) or len(lowercase ) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , lowercase ) is not None def __lowerCAmelCase ( lowercase : Tuple ) -> Any: """simple docstring""" snake_case : Union[str, Any] = object_name.split("." ) snake_case : str = 0 # First let's find the module where our object lives. snake_case : Tuple = parts[i] while i < len(lowercase ) and not os.path.isfile(os.path.join(lowercase , F'{module}.py' ) ): i += 1 if i < len(lowercase ): snake_case : Dict = os.path.join(lowercase , parts[i] ) if i >= len(lowercase ): raise ValueError(F'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(lowercase , F'{module}.py' ) , "r" , encoding="utf-8" , newline="\n" ) as f: snake_case : Any = f.readlines() # Now let's find the class / func in the code! snake_case : Dict = "" snake_case : str = 0 for name in parts[i + 1 :]: while ( line_index < len(lowercase ) and re.search(RF'^{indent}(class|def)\s+{name}(\(|\:)' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowercase ): raise ValueError(F' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). snake_case : Any = line_index while line_index < len(lowercase ) and _should_continue(lines[line_index] , lowercase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 snake_case : str = lines[start_index:line_index] return "".join(lowercase ) __snake_case = re.compile(R"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") __snake_case = re.compile(R"""^\s*(\S+)->(\S+)(\s+.*|$)""") __snake_case = re.compile(R"""<FILL\s+[^>]*>""") def __lowerCAmelCase ( lowercase : str ) -> str: """simple docstring""" snake_case : Dict = code.split("\n" ) snake_case : str = 0 while idx < len(lowercase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowercase ): return re.search(R"^(\s*)\S" , lines[idx] ).groups()[0] return "" def __lowerCAmelCase ( lowercase : List[str] ) -> str: """simple docstring""" snake_case : List[Any] = len(get_indent(lowercase ) ) > 0 if has_indent: snake_case : Union[str, Any] = F'class Bla:\n{code}' snake_case : List[str] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowercase ) snake_case : Any = black.format_str(lowercase , mode=lowercase ) snake_case : Optional[int] = style_docstrings_in_code(lowercase ) return result[len("class Bla:\n" ) :] if has_indent else result def __lowerCAmelCase ( lowercase : Dict , lowercase : Tuple=False ) -> Optional[Any]: """simple docstring""" with open(lowercase , "r" , encoding="utf-8" , newline="\n" ) as f: snake_case : Union[str, Any] = f.readlines() snake_case : str = [] snake_case : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowercase ): snake_case : int = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. snake_case : Any = search.groups() snake_case : Optional[int] = find_code_in_diffusers(lowercase ) snake_case : str = get_indent(lowercase ) snake_case : Any = line_index + 1 if indent == theoretical_indent else line_index + 2 snake_case : str = theoretical_indent snake_case : str = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. snake_case : int = True while line_index < len(lowercase ) and should_continue: line_index += 1 if line_index >= len(lowercase ): break snake_case : List[Any] = lines[line_index] snake_case : Optional[int] = _should_continue(lowercase , lowercase ) and re.search(F'^{indent}# End copy' , lowercase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 snake_case : int = lines[start_index:line_index] snake_case : List[Any] = "".join(lowercase ) # Remove any nested `Copied from` comments to avoid circular copies snake_case : Dict = [line for line in theoretical_code.split("\n" ) if _re_copy_warning.search(lowercase ) is None] snake_case : Tuple = "\n".join(lowercase ) # Before comparing, use the `replace_pattern` on the original code. if len(lowercase ) > 0: snake_case : Union[str, Any] = replace_pattern.replace("with" , "" ).split("," ) snake_case : Tuple = [_re_replace_pattern.search(lowercase ) for p in patterns] for pattern in patterns: if pattern is None: continue snake_case : List[str] = pattern.groups() snake_case : Optional[Any] = re.sub(lowercase , lowercase , lowercase ) if option.strip() == "all-casing": snake_case : Optional[int] = re.sub(obja.lower() , obja.lower() , lowercase ) snake_case : Tuple = re.sub(obja.upper() , obja.upper() , lowercase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line snake_case : Tuple = blackify(lines[start_index - 1] + theoretical_code ) snake_case : Tuple = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: snake_case : Tuple = lines[:start_index] + [theoretical_code] + lines[line_index:] snake_case : List[str] = start_index + 1 if overwrite and len(lowercase ) > 0: # Warn the user a file has been modified. print(F'Detected changes, rewriting {filename}.' ) with open(lowercase , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lowercase ) return diffs def __lowerCAmelCase ( lowercase : bool = False ) -> Dict: """simple docstring""" snake_case : str = glob.glob(os.path.join(lowercase , "**/*.py" ) , recursive=lowercase ) snake_case : Optional[int] = [] for filename in all_files: snake_case : Dict = is_copy_consistent(lowercase , lowercase ) diffs += [F'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(lowercase ) > 0: snake_case : List[Any] = "\n".join(lowercase ) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them." ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") __snake_case = parser.parse_args() check_copies(args.fix_and_overwrite)
371
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
112
0
"""simple docstring""" import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = inspect.getfile(accelerate.test_utils) a__: Any = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['scripts', 'test_script.py']) a__: Dict = os.path.sep.join(inspect.getfile(self.__class__).split(os.path.sep)[:-1]) @require_tpu def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = f'\n {self.test_dir}/xla_spawn.py\n --num_cores 8\n {self.test_file_path}\n '.split() a__: Optional[int] = [sys.executable] + distributed_args execute_subprocess_async(lowercase , env=os.environ.copy())
290
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
1
'''simple docstring''' import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () __SCREAMING_SNAKE_CASE :Tuple = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). __SCREAMING_SNAKE_CASE :Tuple = [0, 25, 50] __SCREAMING_SNAKE_CASE :List[str] = [25, 50, 75] __SCREAMING_SNAKE_CASE :int = fuzz.membership.trimf(X, abca) __SCREAMING_SNAKE_CASE :List[Any] = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. __SCREAMING_SNAKE_CASE :List[str] = np.ones(75) __SCREAMING_SNAKE_CASE :str = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) __SCREAMING_SNAKE_CASE :int = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) __SCREAMING_SNAKE_CASE :List[str] = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) __SCREAMING_SNAKE_CASE :Union[str, Any] = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) __SCREAMING_SNAKE_CASE :List[str] = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] __SCREAMING_SNAKE_CASE :Union[str, Any] = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) __SCREAMING_SNAKE_CASE :Dict = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] __SCREAMING_SNAKE_CASE :int = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] __SCREAMING_SNAKE_CASE :Any = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('''Young''') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('''Middle aged''') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('''union''') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('''intersection''') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('''complement_a''') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('''difference a/b''') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('''alg_sum''') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('''alg_product''') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('''bdd_sum''') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('''bdd_difference''') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
371
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __SCREAMING_SNAKE_CASE :Tuple = '''\ ''' __SCREAMING_SNAKE_CASE :Union[str, Any] = ''' Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity ''' __SCREAMING_SNAKE_CASE :List[Any] = ''' Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to \'cuda\' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"] >>> results = perplexity.compute(model_id=\'gpt2\', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 78.22 >>> print(round(results["perplexities"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = datasets.load_dataset("wikitext", ... "wikitext-2-raw-v1", ... split="test")["text"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=\'\'] >>> results = perplexity.compute(model_id=\'gpt2\', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 60.35 >>> print(round(results["perplexities"][0], 2)) 81.12 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ ( datasets.Metric ): def lowercase ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "input_texts": datasets.Value("string" ), } ) , reference_urls=["https://huggingface.co/docs/transformers/perplexity"] , ) def lowercase ( self : List[Any] , snake_case_ : Optional[Any] , snake_case_ : List[Any] , snake_case_ : int = 1_6 , snake_case_ : bool = True , snake_case_ : int=None ): if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": _UpperCAmelCase = "cuda" else: _UpperCAmelCase = "cuda" if torch.cuda.is_available() else "cpu" _UpperCAmelCase = AutoModelForCausalLM.from_pretrained(snake_case_ ) _UpperCAmelCase = model.to(snake_case_ ) _UpperCAmelCase = AutoTokenizer.from_pretrained(snake_case_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: _UpperCAmelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(snake_case_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"pad_token": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" _UpperCAmelCase = model.config.max_length - 1 else: _UpperCAmelCase = model.config.max_length _UpperCAmelCase = tokenizer( snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , return_tensors="pt" , return_attention_mask=snake_case_ , ).to(snake_case_ ) _UpperCAmelCase = encodings["input_ids"] _UpperCAmelCase = encodings["attention_mask"] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." _UpperCAmelCase = [] _UpperCAmelCase = CrossEntropyLoss(reduction="none" ) for start_index in logging.tqdm(range(0 , len(snake_case_ ) , snake_case_ ) ): _UpperCAmelCase = min(start_index + batch_size , len(snake_case_ ) ) _UpperCAmelCase = encoded_texts[start_index:end_index] _UpperCAmelCase = attn_masks[start_index:end_index] if add_start_token: _UpperCAmelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(snake_case_ ) _UpperCAmelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) _UpperCAmelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(snake_case_ ), attn_mask] , dim=1 ) _UpperCAmelCase = encoded_batch with torch.no_grad(): _UpperCAmelCase = model(snake_case_ , attention_mask=snake_case_ ).logits _UpperCAmelCase = out_logits[..., :-1, :].contiguous() _UpperCAmelCase = labels[..., 1:].contiguous() _UpperCAmelCase = attn_mask[..., 1:].contiguous() _UpperCAmelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , snake_case_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(snake_case_ )}
156
0
'''simple docstring''' import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize("""dataset_size""" , [None, 4_0_0 * 2**2_0, 6_0_0 * 2**2_0] ) @pytest.mark.parametrize("""input_in_memory_max_size""" , ["""default""", 0, 1_0_0 * 2**2_0, 9_0_0 * 2**2_0] ) def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> Union[str, Any]: if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , """IN_MEMORY_MAX_SIZE""" , lowerCAmelCase__ ) UpperCAmelCase : Optional[Any] = datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: UpperCAmelCase : int = dataset_size < in_memory_max_size else: UpperCAmelCase : str = False UpperCAmelCase : Union[str, Any] = is_small_dataset(lowerCAmelCase__ ) assert result == expected
265
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Union[str, Any] = logging.get_logger(__name__) lowercase__ : List[str] = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCamelCase__ ( lowercase_ ): """simple docstring""" _SCREAMING_SNAKE_CASE = """informer""" _SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : Any , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : str = "student_t" , SCREAMING_SNAKE_CASE_ : str = "nll" , SCREAMING_SNAKE_CASE_ : int = 1 , SCREAMING_SNAKE_CASE_ : List[int] = None , SCREAMING_SNAKE_CASE_ : Optional[Union[str, bool]] = "mean" , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : int = 6_4 , SCREAMING_SNAKE_CASE_ : int = 3_2 , SCREAMING_SNAKE_CASE_ : int = 3_2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : bool = True , SCREAMING_SNAKE_CASE_ : str = "gelu" , SCREAMING_SNAKE_CASE_ : float = 0.05 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : int = 1_0_0 , SCREAMING_SNAKE_CASE_ : float = 0.02 , SCREAMING_SNAKE_CASE_ : str=True , SCREAMING_SNAKE_CASE_ : str = "prob" , SCREAMING_SNAKE_CASE_ : int = 5 , SCREAMING_SNAKE_CASE_ : bool = True , **SCREAMING_SNAKE_CASE_ : int , ): # time series specific configuration lowerCAmelCase_ : Dict = prediction_length lowerCAmelCase_ : List[str] = context_length or prediction_length lowerCAmelCase_ : List[Any] = distribution_output lowerCAmelCase_ : int = loss lowerCAmelCase_ : Optional[int] = input_size lowerCAmelCase_ : Tuple = num_time_features lowerCAmelCase_ : List[str] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] lowerCAmelCase_ : int = scaling lowerCAmelCase_ : List[Any] = num_dynamic_real_features lowerCAmelCase_ : Union[str, Any] = num_static_real_features lowerCAmelCase_ : Optional[int] = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(SCREAMING_SNAKE_CASE_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) lowerCAmelCase_ : str = cardinality else: lowerCAmelCase_ : Any = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(SCREAMING_SNAKE_CASE_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) lowerCAmelCase_ : Optional[int] = embedding_dimension else: lowerCAmelCase_ : Union[str, Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCAmelCase_ : Optional[int] = num_parallel_samples # Transformer architecture configuration lowerCAmelCase_ : Any = input_size * len(self.lags_sequence ) + self._number_of_features lowerCAmelCase_ : Any = d_model lowerCAmelCase_ : Union[str, Any] = encoder_attention_heads lowerCAmelCase_ : Optional[Any] = decoder_attention_heads lowerCAmelCase_ : Any = encoder_ffn_dim lowerCAmelCase_ : List[str] = decoder_ffn_dim lowerCAmelCase_ : Optional[Any] = encoder_layers lowerCAmelCase_ : Tuple = decoder_layers lowerCAmelCase_ : Optional[int] = dropout lowerCAmelCase_ : Dict = attention_dropout lowerCAmelCase_ : int = activation_dropout lowerCAmelCase_ : Dict = encoder_layerdrop lowerCAmelCase_ : str = decoder_layerdrop lowerCAmelCase_ : Union[str, Any] = activation_function lowerCAmelCase_ : Union[str, Any] = init_std lowerCAmelCase_ : Union[str, Any] = use_cache # Informer lowerCAmelCase_ : Optional[int] = attention_type lowerCAmelCase_ : Any = sampling_factor lowerCAmelCase_ : int = distil super().__init__(is_encoder_decoder=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) @property def SCREAMING_SNAKE_CASE__ ( self : Any ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
224
0
import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class __A ( unittest.TestCase ): def __init__( self , UpperCAmelCase_ ): lowerCamelCase =parent def _snake_case ( self ): return {} def _lowercase ( ) -> Any: lowerCamelCase ="""<HTML> <HEAD> <TITLE>sample document</TITLE> </HEAD> <BODY BGCOLOR=\"FFFFFF\"> <HR> <a href=\"http://google.com\">Goog</a> <H1>This is one header</H1> <H2>This is a another Header</H2> <P>Travel from <P> <B>SFO to JFK</B> <BR> <B><I>on May 2, 2015 at 2:00 pm. For details go to confirm.com </I></B> <HR> <div style=\"color:#0000FF\"> <h3>Traveler <b> name </b> is <p> John Doe </p> </div>""" lowerCamelCase =""" <!DOCTYPE html> <html> <body> <h1>My First Heading</h1> <p>My first paragraph.</p> </body> </html> """ return [html_string_a, html_string_a] @require_bsa class __A ( a , unittest.TestCase ): __A = MarkupLMFeatureExtractor if is_bsa_available() else None def _snake_case ( self ): lowerCamelCase =MarkupLMFeatureExtractionTester(self ) @property def _snake_case ( self ): return self.feature_extract_tester.prepare_feat_extract_dict() def _snake_case ( self ): # Initialize feature_extractor lowerCamelCase =self.feature_extraction_class() # Test not batched input lowerCamelCase =get_html_strings()[0] lowerCamelCase =feature_extractor(UpperCAmelCase_ ) # fmt: off lowerCamelCase =[["""sample document""", """Goog""", """This is one header""", """This is a another Header""", """Travel from""", """SFO to JFK""", """on May 2, 2015 at 2:00 pm. For details go to confirm.com""", """Traveler""", """name""", """is""", """John Doe"""]] lowerCamelCase =[["""/html/head/title""", """/html/body/a""", """/html/body/h1""", """/html/body/h2""", """/html/body/p""", """/html/body/p/p/b[1]""", """/html/body/p/p/b[2]/i""", """/html/body/p/p/div/h3""", """/html/body/p/p/div/h3/b""", """/html/body/p/p/div/h3""", """/html/body/p/p/div/h3/p"""]] # fmt: on self.assertEqual(encoding.nodes , UpperCAmelCase_ ) self.assertEqual(encoding.xpaths , UpperCAmelCase_ ) # Test batched lowerCamelCase =get_html_strings() lowerCamelCase =feature_extractor(UpperCAmelCase_ ) # fmt: off lowerCamelCase =expected_nodes + [["""My First Heading""", """My first paragraph."""]] lowerCamelCase =expected_xpaths + [["""/html/body/h1""", """/html/body/p"""]] self.assertEqual(len(encoding.nodes ) , 2 ) self.assertEqual(len(encoding.xpaths ) , 2 ) self.assertEqual(encoding.nodes , UpperCAmelCase_ ) self.assertEqual(encoding.xpaths , UpperCAmelCase_ )
262
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class __A ( a , unittest.TestCase ): __A = BioGptTokenizer __A = False def _snake_case ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCamelCase =[ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] lowerCamelCase =dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCamelCase =["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] lowerCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(UpperCAmelCase_ ) ) def _snake_case ( self , UpperCAmelCase_ ): lowerCamelCase ="""lower newer""" lowerCamelCase ="""lower newer""" return input_text, output_text def _snake_case ( self ): lowerCamelCase =BioGptTokenizer(self.vocab_file , self.merges_file ) lowerCamelCase ="""lower""" lowerCamelCase =["""low""", """er</w>"""] lowerCamelCase =tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCamelCase =tokens + ["""<unk>"""] lowerCamelCase =[14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , UpperCAmelCase_ ) @slow def _snake_case ( self ): lowerCamelCase =BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) lowerCamelCase =tokenizer.encode("""sequence builders""" , add_special_tokens=UpperCAmelCase_ ) lowerCamelCase =tokenizer.encode("""multi-sequence build""" , add_special_tokens=UpperCAmelCase_ ) lowerCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
262
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowercase : str = logging.get_logger(__name__) lowercase : Dict = '▁' lowercase : Optional[Any] = {'vocab_file': 'sentencepiece.bpe.model'} lowercase : Union[str, Any] = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } lowercase : Union[str, Any] = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off lowercase : Optional[Any] = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class lowerCamelCase__ ( _A): '''simple docstring''' _A = VOCAB_FILES_NAMES _A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A = PRETRAINED_VOCAB_FILES_MAP _A = ['input_ids', 'attention_mask'] _A = [] _A = [] def __init__( self :List[str] , a :Tuple , a :Tuple="<s>" , a :int="</s>" , a :Any="</s>" , a :int="<s>" , a :Dict="<unk>" , a :Tuple="<pad>" , a :str="<mask>" , a :Dict=None , a :str=None , a :Any=None , a :Optional[Dict[str, Any]] = None , a :Optional[int]=None , **a :int , ) -> List[Any]: __UpperCamelCase : Union[str, Any] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token __UpperCamelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , src_lang=UpperCamelCase__ , tgt_lang=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) __UpperCamelCase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase__ ) ) __UpperCamelCase : Optional[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token __UpperCamelCase : Any = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __UpperCamelCase : int = 1 __UpperCamelCase : int = len(self.sp_model ) __UpperCamelCase : Optional[int] = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(UpperCamelCase__ ) } __UpperCamelCase : Optional[Any] = {v: k for k, v in self.lang_code_to_id.items()} __UpperCamelCase : int = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) __UpperCamelCase : str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} __UpperCamelCase : List[Any] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) __UpperCamelCase : str = src_lang if src_lang is not None else "en_XX" __UpperCamelCase : Dict = self.lang_code_to_id[self._src_lang] __UpperCamelCase : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self :int ) -> int: __UpperCamelCase : str = self.__dict__.copy() __UpperCamelCase : Union[str, Any] = None __UpperCamelCase : str = self.sp_model.serialized_model_proto() return state def __setstate__( self :Tuple , a :int ) -> Dict: __UpperCamelCase : List[str] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __UpperCamelCase : Tuple = {} __UpperCamelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def _lowerCamelCase ( self :Dict ) -> str: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def _lowerCamelCase ( self :Union[str, Any] ) -> str: return self._src_lang @src_lang.setter def _lowerCamelCase ( self :Optional[int] , a :str ) -> None: __UpperCamelCase : Dict = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCamelCase ( self :Union[str, Any] , a :List[int] , a :Optional[List[int]] = None , a :bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) __UpperCamelCase : Optional[Any] = [1] * len(self.prefix_tokens ) __UpperCamelCase : Dict = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(UpperCamelCase__ )) + suffix_ones return prefix_ones + ([0] * len(UpperCamelCase__ )) + ([0] * len(UpperCamelCase__ )) + suffix_ones def _lowerCamelCase ( self :List[Any] , a :List[int] , a :Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCamelCase ( self :Union[str, Any] , a :List[int] , a :Optional[List[int]] = None ) -> List[int]: __UpperCamelCase : Union[str, Any] = [self.sep_token_id] __UpperCamelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowerCamelCase ( self :Optional[Any] , a :Tuple , a :str , a :Optional[str] , a :Optional[str] , **a :int ) -> List[Any]: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) __UpperCamelCase : Optional[int] = src_lang __UpperCamelCase : Tuple = self(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase : Dict = self.convert_tokens_to_ids(UpperCamelCase__ ) __UpperCamelCase : List[str] = tgt_lang_id return inputs def _lowerCamelCase ( self :Optional[int] ) -> Any: __UpperCamelCase : List[str] = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowerCamelCase ( self :Optional[Any] , a :str ) -> List[str]: return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def _lowerCamelCase ( self :Dict , a :Optional[int] ) -> List[str]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __UpperCamelCase : Optional[Any] = self.sp_model.PieceToId(UpperCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _lowerCamelCase ( self :Any , a :List[Any] ) -> Any: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _lowerCamelCase ( self :Any , a :Tuple ) -> str: __UpperCamelCase : Dict = "".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _lowerCamelCase ( self :Dict , a :str , a :Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCamelCase__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __UpperCamelCase : List[Any] = os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: __UpperCamelCase : Dict = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def _lowerCamelCase ( self :List[str] , a :List[str] , a :str = "en_XX" , a :Optional[List[str]] = None , a :str = "ro_RO" , **a :int , ) -> BatchEncoding: __UpperCamelCase : List[Any] = src_lang __UpperCamelCase : Any = tgt_lang return super().prepare_seqaseq_batch(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def _lowerCamelCase ( self :Optional[int] ) -> List[str]: return self.set_src_lang_special_tokens(self.src_lang ) def _lowerCamelCase ( self :Optional[Any] ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCamelCase ( self :Union[str, Any] , a :str ) -> None: __UpperCamelCase : Union[str, Any] = self.lang_code_to_id[src_lang] __UpperCamelCase : Any = [] __UpperCamelCase : Dict = [self.eos_token_id, self.cur_lang_code] def _lowerCamelCase ( self :Tuple , a :str ) -> None: __UpperCamelCase : Optional[Any] = self.lang_code_to_id[lang] __UpperCamelCase : str = [] __UpperCamelCase : Union[str, Any] = [self.eos_token_id, self.cur_lang_code]
232
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger __lowerCAmelCase : Any = get_logger(__name__) class UpperCAmelCase_ : '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Optional[str] = None ) -> Optional[Any]: """simple docstring""" __magic_name__ = ( os.path.join(UpperCamelCase__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) __magic_name__ = Extractor def _lowercase ( self : Optional[Any] , UpperCamelCase__ : str ) -> str: """simple docstring""" from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" __magic_name__ = os.path.abspath(UpperCamelCase__ ) return os.path.join(self.extract_dir , hash_url_to_filename(UpperCamelCase__ ) ) def _lowercase ( self : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : bool ) -> bool: """simple docstring""" return force_extract or ( not os.path.isfile(UpperCamelCase__ ) and not (os.path.isdir(UpperCamelCase__ ) and os.listdir(UpperCamelCase__ )) ) def _lowercase ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : bool = False ) -> str: """simple docstring""" __magic_name__ = self.extractor.infer_extractor_format(UpperCamelCase__ ) if not extractor_format: return input_path __magic_name__ = self._get_output_path(UpperCamelCase__ ) if self._do_extract(UpperCamelCase__ , UpperCamelCase__ ): self.extractor.extract(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return output_path class UpperCAmelCase_ ( _A ): '''simple docstring''' @classmethod @abstractmethod def _lowercase ( cls : List[str] , UpperCamelCase__ : Union[Path, str] , **UpperCamelCase__ : Union[str, Any] ) -> bool: """simple docstring""" ... @staticmethod @abstractmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" ... class UpperCAmelCase_ ( _A , _A ): '''simple docstring''' a__ = [] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : int ) -> List[str]: """simple docstring""" with open(UpperCamelCase__ , """rb""" ) as f: return f.read(UpperCamelCase__ ) @classmethod def _lowercase ( cls : List[Any] , UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : bytes = b"" ) -> bool: """simple docstring""" if not magic_number: __magic_name__ = max(len(UpperCamelCase__ ) for cls_magic_number in cls.magic_numbers ) try: __magic_name__ = cls.read_magic_number(UpperCamelCase__ , UpperCamelCase__ ) except OSError: return False return any(magic_number.startswith(UpperCamelCase__ ) for cls_magic_number in cls.magic_numbers ) class UpperCAmelCase_ ( _A ): '''simple docstring''' @classmethod def _lowercase ( cls : Optional[Any] , UpperCamelCase__ : Union[Path, str] , **UpperCamelCase__ : int ) -> bool: """simple docstring""" return tarfile.is_tarfile(UpperCamelCase__ ) @staticmethod def _lowercase ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[int] ) -> Union[str, Any]: """simple docstring""" def resolved(UpperCamelCase__ : str ) -> str: return os.path.realpath(os.path.abspath(UpperCamelCase__ ) ) def badpath(UpperCamelCase__ : str , UpperCamelCase__ : str ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ).startswith(UpperCamelCase__ ) def badlink(UpperCamelCase__ : Optional[int] , UpperCamelCase__ : str ) -> bool: # Links are interpreted relative to the directory containing the link __magic_name__ = resolved(os.path.join(UpperCamelCase__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=UpperCamelCase__ ) __magic_name__ = resolved(UpperCamelCase__ ) for finfo in members: if badpath(finfo.name , UpperCamelCase__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(UpperCamelCase__ , UpperCamelCase__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(UpperCamelCase__ , UpperCamelCase__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) __magic_name__ = tarfile.open(UpperCamelCase__ ) tar_file.extractall(UpperCamelCase__ , members=TarExtractor.safemembers(UpperCamelCase__ , UpperCamelCase__ ) ) tar_file.close() class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\x1F\x8B"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" with gzip.open(UpperCamelCase__ , """rb""" ) as gzip_file: with open(UpperCamelCase__ , """wb""" ) as extracted_file: shutil.copyfileobj(UpperCamelCase__ , UpperCamelCase__ ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def _lowercase ( cls : Union[str, Any] , UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : bytes = b"" ) -> bool: """simple docstring""" if super().is_extractable(UpperCamelCase__ , magic_number=UpperCamelCase__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(UpperCamelCase__ , """rb""" ) as fp: __magic_name__ = _EndRecData(UpperCamelCase__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: __magic_name__ = fp.read(UpperCamelCase__ ) # CD is where we expect it to be if len(UpperCamelCase__ ) == sizeCentralDir: __magic_name__ = struct.unpack(UpperCamelCase__ , UpperCamelCase__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) with zipfile.ZipFile(UpperCamelCase__ , """r""" ) as zip_file: zip_file.extractall(UpperCamelCase__ ) zip_file.close() class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" with lzma.open(UpperCamelCase__ ) as compressed_file: with open(UpperCamelCase__ , """wb""" ) as extracted_file: shutil.copyfileobj(UpperCamelCase__ , UpperCamelCase__ ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) __magic_name__ = rarfile.RarFile(UpperCamelCase__ ) rf.extractall(UpperCamelCase__ ) rf.close() class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd __magic_name__ = zstd.ZstdDecompressor() with open(UpperCamelCase__ , """rb""" ) as ifh, open(UpperCamelCase__ , """wb""" ) as ofh: dctx.copy_stream(UpperCamelCase__ , UpperCamelCase__ ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\x42\x5A\x68"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" with bza.open(UpperCamelCase__ , """rb""" ) as compressed_file: with open(UpperCamelCase__ , """wb""" ) as extracted_file: shutil.copyfileobj(UpperCamelCase__ , UpperCamelCase__ ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) with pyazr.SevenZipFile(UpperCamelCase__ , """r""" ) as archive: archive.extractall(UpperCamelCase__ ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = [B"""\x04\x22\x4D\x18"""] @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] ) -> None: """simple docstring""" if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(UpperCamelCase__ , """rb""" ) as compressed_file: with open(UpperCamelCase__ , """wb""" ) as extracted_file: shutil.copyfileobj(UpperCamelCase__ , UpperCamelCase__ ) class UpperCAmelCase_ : '''simple docstring''' a__ = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def _lowercase ( cls : Tuple ) -> Tuple: """simple docstring""" return max( len(UpperCamelCase__ ) for extractor in cls.extractors.values() if issubclass(UpperCamelCase__ , UpperCamelCase__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def _lowercase ( UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : int ) -> Union[str, Any]: """simple docstring""" try: return MagicNumberBaseExtractor.read_magic_number(UpperCamelCase__ , magic_number_length=UpperCamelCase__ ) except OSError: return b"" @classmethod def _lowercase ( cls : List[Any] , UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : bool = False ) -> bool: """simple docstring""" warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=UpperCamelCase__ , ) __magic_name__ = cls.infer_extractor_format(UpperCamelCase__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def _lowercase ( cls : Dict , UpperCamelCase__ : Union[Path, str] ) -> str: # <Added version="2.4.0"/> """simple docstring""" __magic_name__ = cls._get_magic_number_max_length() __magic_name__ = cls._read_magic_number(UpperCamelCase__ , UpperCamelCase__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(UpperCamelCase__ , magic_number=UpperCamelCase__ ): return extractor_format @classmethod def _lowercase ( cls : Union[str, Any] , UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Union[Path, str] , UpperCamelCase__ : Optional[str] = None , UpperCamelCase__ : Optional[BaseExtractor] = "deprecated" , ) -> None: """simple docstring""" os.makedirs(os.path.dirname(UpperCamelCase__ ) , exist_ok=UpperCamelCase__ ) # Prevent parallel extractions __magic_name__ = str(Path(UpperCamelCase__ ).with_suffix(""".lock""" ) ) with FileLock(UpperCamelCase__ ): shutil.rmtree(UpperCamelCase__ , ignore_errors=UpperCamelCase__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(UpperCamelCase__ , UpperCamelCase__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=UpperCamelCase__ , ) __magic_name__ = extractor if extractor != """deprecated""" else extractor_format else: __magic_name__ = cls.extractors[extractor_format] return extractor.extract(UpperCamelCase__ , UpperCamelCase__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=UpperCamelCase__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(UpperCamelCase__ ): return extractor.extract(UpperCamelCase__ , UpperCamelCase__ )
88
0
def lowerCAmelCase_ ( snake_case_ ): if not isinstance(snake_case_,snake_case_ ): _A : Any = f'''Input value of [number={number}] must be an integer''' raise TypeError(snake_case_ ) if number < 0: return False _A : List[str] = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
343
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _snake_case = logging.getLogger() def lowerCAmelCase_ ( ): _A : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""-f""" ) _A : Optional[Any] = parser.parse_args() return args.f class lowercase ( UpperCamelCase__ ): def a__ ( self ) -> None: _A : List[Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(_a ) def a__ ( self , _a ) -> Dict: _A : Tuple = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , """run_glue_deebert.py""" ) with patch.object(_a , """argv""" , _a ): _A : Optional[Any] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(_a , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ) -> Optional[int]: _A : Tuple = """ --model_type roberta --model_name_or_path roberta-base --task_name MRPC --do_train --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --max_seq_length 128 --per_gpu_eval_batch_size=1 --per_gpu_train_batch_size=8 --learning_rate 2e-4 --num_train_epochs 3 --overwrite_output_dir --seed 42 --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --save_steps 0 --overwrite_cache --eval_after_first_stage """.split() self.run_and_check(_a ) _A : Optional[Any] = """ --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --eval_each_highway --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 """.split() self.run_and_check(_a ) _A : List[str] = """ --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --early_exit_entropy 0.1 --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 """.split() self.run_and_check(_a )
343
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class __lowerCAmelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=2 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=1_0 , lowerCAmelCase__=3 , lowerCAmelCase__=3_2 * 8 , lowerCAmelCase__=3_2 * 8 , lowerCAmelCase__=4 , lowerCAmelCase__=6_4 , ) -> List[str]: '''simple docstring''' a__ : int =parent a__ : List[Any] =batch_size a__ : int =is_training a__ : Optional[Any] =use_auxiliary_loss a__ : int =num_queries a__ : Any =num_channels a__ : Any =min_size a__ : Optional[Any] =max_size a__ : List[str] =num_labels a__ : Dict =hidden_dim a__ : Dict =hidden_dim def _lowercase ( self ) -> List[Any]: '''simple docstring''' a__ : List[Any] =floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowerCAmelCase__ ) a__ : Tuple =torch.ones([self.batch_size, self.min_size, self.max_size] , device=lowerCAmelCase__ ) a__ : Dict =( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=lowerCAmelCase__ ) > 0.5 ).float() a__ : Dict =(torch.rand((self.batch_size, self.num_labels) , device=lowerCAmelCase__ ) > 0.5).long() a__ : str =self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _lowercase ( self ) -> Dict: '''simple docstring''' a__ : Optional[Any] =MaskaFormerConfig( hidden_size=self.hidden_dim , ) a__ : Optional[Any] =self.num_queries a__ : int =self.num_labels a__ : Optional[Any] =[1, 1, 1, 1] a__ : Dict =self.num_channels a__ : List[str] =6_4 a__ : Tuple =1_2_8 a__ : List[str] =self.hidden_dim a__ : str =self.hidden_dim a__ : int =self.hidden_dim return config def _lowercase ( self ) -> List[Any]: '''simple docstring''' a__ , a__ , a__ , a__ , a__ : int =self.prepare_config_and_inputs() a__ : Any ={"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: '''simple docstring''' a__ : Optional[Any] =output.encoder_hidden_states a__ : str =output.pixel_decoder_hidden_states a__ : Dict =output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , config.decoder_layers ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ) -> List[Any]: '''simple docstring''' with torch.no_grad(): a__ : str =MaskaFormerModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a__ : Union[str, Any] =model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) a__ : Optional[int] =model(lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: '''simple docstring''' a__ : Any =MaskaFormerForUniversalSegmentation(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() def comm_check_on_output(lowerCAmelCase__ ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): a__ : Optional[Any] =model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) a__ : Any =model(lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) a__ : Optional[int] =model( pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): _lowercase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () _lowercase : Union[str, Any] = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} _lowercase : str = False _lowercase : List[Any] = False _lowercase : List[str] = False _lowercase : int = False def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : Optional[Any] =MaskaFormerModelTester(self ) a__ : Optional[Any] =ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ) -> int: '''simple docstring''' a__ , a__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' a__ : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*lowerCAmelCase__ ) @unittest.skip(reason="Mask2Former does not use inputs_embeds" ) def _lowercase ( self ) -> str: '''simple docstring''' pass @unittest.skip(reason="Mask2Former does not have a get_input_embeddings method" ) def _lowercase ( self ) -> int: '''simple docstring''' pass @unittest.skip(reason="Mask2Former is not a generative model" ) def _lowercase ( self ) -> Tuple: '''simple docstring''' pass @unittest.skip(reason="Mask2Former does not use token embeddings" ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def _lowercase ( self ) -> Any: '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _lowercase ( self ) -> Tuple: '''simple docstring''' pass def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ , a__ : int =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : Dict =model_class(lowerCAmelCase__ ) a__ : List[str] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__ : Dict =[*signature.parameters.keys()] a__ : List[str] =["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) @slow def _lowercase ( self ) -> Any: '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: a__ : Optional[int] =MaskaFormerModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _lowercase ( self ) -> str: '''simple docstring''' a__ : List[str] =(self.model_tester.min_size,) * 2 a__ : Any ={ "pixel_values": torch.randn((2, 3, *size) , device=lowerCAmelCase__ ), "mask_labels": torch.randn((2, 1_0, *size) , device=lowerCAmelCase__ ), "class_labels": torch.zeros(2 , 1_0 , device=lowerCAmelCase__ ).long(), } a__ : Any =self.model_tester.get_config() a__ : str =MaskaFormerForUniversalSegmentation(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a__ : Any =model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None ) def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ , a__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ , a__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : List[Any] =model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a__ : Optional[int] =model(**lowerCAmelCase__ , output_attentions=lowerCAmelCase__ ) self.assertTrue(outputs.attentions is not None ) def _lowercase ( self ) -> Tuple: '''simple docstring''' if not self.model_tester.is_training: return a__ : Optional[Any] =self.all_model_classes[1] a__ , a__ , a__ , a__ , a__ : Optional[int] =self.model_tester.prepare_config_and_inputs() a__ : str =model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() a__ : List[str] =model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ).loss loss.backward() def _lowercase ( self ) -> str: '''simple docstring''' a__ : Dict =self.all_model_classes[1] a__ , a__ , a__ , a__ , a__ : Dict =self.model_tester.prepare_config_and_inputs() a__ : Tuple =True a__ : int =True a__ : List[Any] =model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) model.train() a__ : List[str] =model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) a__ : Optional[int] =outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() a__ : Optional[Any] =outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() a__ : Tuple =outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() a__ : Optional[Any] =outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowerCAmelCase__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) UpperCAmelCase : int = 1E-4 def _A ( ): """simple docstring""" a__ : int =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class __lowerCAmelCase ( unittest.TestCase): @cached_property def _lowercase ( self ) -> List[str]: '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : Optional[int] =MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ) a__ : Tuple =self.default_image_processor a__ : Tuple =prepare_img() a__ : int =image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) a__ : List[Any] =inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 3_8_4, 3_8_4) ) with torch.no_grad(): a__ : Tuple =model(**lowerCAmelCase__ ) a__ : Union[str, Any] =torch.tensor( [[-0.27_90, -1.07_17, -1.16_68], [-0.51_28, -0.31_28, -0.49_87], [-0.58_32, 0.19_71, -0.01_97]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) a__ : List[Any] =torch.tensor( [[0.89_73, 1.18_47, 1.17_76], [1.19_34, 1.50_40, 1.51_28], [1.11_53, 1.44_86, 1.49_51]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) a__ : str =torch.tensor( [[2.11_52, 1.70_00, -0.86_03], [1.58_08, 1.80_04, -0.93_53], [1.60_43, 1.74_95, -0.59_99]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _lowercase ( self ) -> str: '''simple docstring''' a__ : Optional[int] =MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ).eval() a__ : List[str] =self.default_image_processor a__ : Optional[int] =prepare_img() a__ : Optional[int] =image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) a__ : Union[str, Any] =inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 3_8_4, 3_8_4) ) with torch.no_grad(): a__ : Tuple =model(**lowerCAmelCase__ ) # masks_queries_logits a__ : Tuple =outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) a__ : int =[ [-8.78_39, -9.00_56, -8.81_21], [-7.41_04, -7.03_13, -6.54_01], [-6.61_05, -6.34_27, -6.46_75], ] a__ : List[str] =torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) # class_queries_logits a__ : str =outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) a__ : Any =torch.tensor( [ [1.83_24, -8.08_35, -4.19_22], [0.84_50, -9.00_50, -3.60_53], [0.30_45, -7.72_93, -3.02_75], ] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : Union[str, Any] =MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ).eval() a__ : Union[str, Any] =self.default_image_processor a__ : Dict =image_processor( [np.zeros((3, 8_0_0, 1_3_3_3) ), np.zeros((3, 8_0_0, 1_3_3_3) )] , segmentation_maps=[np.zeros((3_8_4, 3_8_4) ).astype(np.floataa ), np.zeros((3_8_4, 3_8_4) ).astype(np.floataa )] , return_tensors="pt" , ) a__ : int =inputs["pixel_values"].to(lowerCAmelCase__ ) a__ : str =[el.to(lowerCAmelCase__ ) for el in inputs["mask_labels"]] a__ : Any =[el.to(lowerCAmelCase__ ) for el in inputs["class_labels"]] with torch.no_grad(): a__ : Tuple =model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None )
95
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class lowercase_ ( __SCREAMING_SNAKE_CASE ): A__ : torch.FloatTensor class lowercase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): @register_to_config def __init__( self , __UpperCamelCase = 3 , __UpperCamelCase = 3 , __UpperCamelCase = ("DownEncoderBlock2D",) , __UpperCamelCase = ("UpDecoderBlock2D",) , __UpperCamelCase = (6_4,) , __UpperCamelCase = 1 , __UpperCamelCase = "silu" , __UpperCamelCase = 3 , __UpperCamelCase = 3_2 , __UpperCamelCase = 2_5_6 , __UpperCamelCase = 3_2 , __UpperCamelCase = None , __UpperCamelCase = 0.18_215 , __UpperCamelCase = "group" , ): """simple docstring""" super().__init__() # pass init params to Encoder UpperCamelCase_ = Encoder( in_channels=__UpperCamelCase , out_channels=__UpperCamelCase , down_block_types=__UpperCamelCase , block_out_channels=__UpperCamelCase , layers_per_block=__UpperCamelCase , act_fn=__UpperCamelCase , norm_num_groups=__UpperCamelCase , double_z=__UpperCamelCase , ) UpperCamelCase_ = vq_embed_dim if vq_embed_dim is not None else latent_channels UpperCamelCase_ = nn.Convad(__UpperCamelCase , __UpperCamelCase , 1 ) UpperCamelCase_ = VectorQuantizer(__UpperCamelCase , __UpperCamelCase , beta=0.25 , remap=__UpperCamelCase , sane_index_shape=__UpperCamelCase ) UpperCamelCase_ = nn.Convad(__UpperCamelCase , __UpperCamelCase , 1 ) # pass init params to Decoder UpperCamelCase_ = Decoder( in_channels=__UpperCamelCase , out_channels=__UpperCamelCase , up_block_types=__UpperCamelCase , block_out_channels=__UpperCamelCase , layers_per_block=__UpperCamelCase , act_fn=__UpperCamelCase , norm_num_groups=__UpperCamelCase , norm_type=__UpperCamelCase , ) @apply_forward_hook def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = True ): """simple docstring""" UpperCamelCase_ = self.encoder(__UpperCamelCase ) UpperCamelCase_ = self.quant_conv(__UpperCamelCase ) if not return_dict: return (h,) return VQEncoderOutput(latents=__UpperCamelCase ) @apply_forward_hook def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = False , __UpperCamelCase = True ): """simple docstring""" if not force_not_quantize: UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = self.quantize(__UpperCamelCase ) else: UpperCamelCase_ = h UpperCamelCase_ = self.post_quant_conv(__UpperCamelCase ) UpperCamelCase_ = self.decoder(__UpperCamelCase , quant if self.config.norm_type == """spatial""" else None ) if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = True ): """simple docstring""" UpperCamelCase_ = sample UpperCamelCase_ = self.encode(__UpperCamelCase ).latents UpperCamelCase_ = self.decode(__UpperCamelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCamelCase )
122
0
'''simple docstring''' import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = AudioLDMPipeline SCREAMING_SNAKE_CASE : Dict = TEXT_TO_AUDIO_PARAMS SCREAMING_SNAKE_CASE : Union[str, Any] = TEXT_TO_AUDIO_BATCH_PARAMS SCREAMING_SNAKE_CASE : str = frozenset( [ 'num_inference_steps', 'num_waveforms_per_prompt', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) def SCREAMING_SNAKE_CASE ( self : int ): torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) ,layers_per_block=2 ,sample_size=3_2 ,in_channels=4 ,out_channels=4 ,down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') ,up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') ,cross_attention_dim=(3_2, 6_4) ,class_embed_type='''simple_projection''' ,projection_class_embeddings_input_dim=3_2 ,class_embeddings_concat=lowercase__ ,) __lowercase = DDIMScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule='''scaled_linear''' ,clip_sample=lowercase__ ,set_alpha_to_one=lowercase__ ,) torch.manual_seed(0 ) __lowercase = AutoencoderKL( block_out_channels=[3_2, 6_4] ,in_channels=1 ,out_channels=1 ,down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] ,up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] ,latent_channels=4 ,) torch.manual_seed(0 ) __lowercase = ClapTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=3_2 ,intermediate_size=3_7 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_0_0_0 ,projection_dim=3_2 ,) __lowercase = ClapTextModelWithProjection(lowercase__ ) __lowercase = RobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-random-roberta''' ,model_max_length=7_7 ) __lowercase = SpeechTaHifiGanConfig( model_in_dim=8 ,sampling_rate=1_6_0_0_0 ,upsample_initial_channel=1_6 ,upsample_rates=[2, 2] ,upsample_kernel_sizes=[4, 4] ,resblock_kernel_sizes=[3, 7] ,resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] ,normalize_before=lowercase__ ,) __lowercase = SpeechTaHifiGan(lowercase__ ) __lowercase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''vocoder''': vocoder, } return components def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Union[str, Any]=0 ): if str(lowercase__ ).startswith('''mps''' ): __lowercase = torch.manual_seed(lowercase__ ) else: __lowercase = torch.Generator(device=lowercase__ ).manual_seed(lowercase__ ) __lowercase = { '''prompt''': '''A hammer hitting a wooden surface''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, } return inputs def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowercase = self.get_dummy_components() __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = audioldm_pipe(**lowercase__ ) __lowercase = output.audios[0] assert audio.ndim == 1 assert len(lowercase__ ) == 2_5_6 __lowercase = audio[:1_0] __lowercase = np.array( [-0.0_0_5_0, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_3, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_3] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.get_dummy_components() __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = 3 * [inputs['''prompt''']] # forward __lowercase = audioldm_pipe(**lowercase__ ) __lowercase = output.audios[0] __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = 3 * [inputs.pop('''prompt''' )] __lowercase = audioldm_pipe.tokenizer( lowercase__ ,padding='''max_length''' ,max_length=audioldm_pipe.tokenizer.model_max_length ,truncation=lowercase__ ,return_tensors='''pt''' ,) __lowercase = text_inputs['''input_ids'''].to(lowercase__ ) __lowercase = audioldm_pipe.text_encoder( lowercase__ ,) __lowercase = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state __lowercase = F.normalize(lowercase__ ,dim=-1 ) __lowercase = prompt_embeds # forward __lowercase = audioldm_pipe(**lowercase__ ) __lowercase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.get_dummy_components() __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = 3 * ['''this is a negative prompt'''] __lowercase = negative_prompt __lowercase = 3 * [inputs['''prompt''']] # forward __lowercase = audioldm_pipe(**lowercase__ ) __lowercase = output.audios[0] __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = 3 * [inputs.pop('''prompt''' )] __lowercase = [] for p in [prompt, negative_prompt]: __lowercase = audioldm_pipe.tokenizer( lowercase__ ,padding='''max_length''' ,max_length=audioldm_pipe.tokenizer.model_max_length ,truncation=lowercase__ ,return_tensors='''pt''' ,) __lowercase = text_inputs['''input_ids'''].to(lowercase__ ) __lowercase = audioldm_pipe.text_encoder( lowercase__ ,) __lowercase = text_embeds.text_embeds # additional L_2 normalization over each hidden-state __lowercase = F.normalize(lowercase__ ,dim=-1 ) embeds.append(lowercase__ ) __lowercase , __lowercase = embeds # forward __lowercase = audioldm_pipe(**lowercase__ ) __lowercase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowercase = self.get_dummy_components() __lowercase = PNDMScheduler(skip_prk_steps=lowercase__ ) __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = '''egg cracking''' __lowercase = audioldm_pipe(**lowercase__ ,negative_prompt=lowercase__ ) __lowercase = output.audios[0] assert audio.ndim == 1 assert len(lowercase__ ) == 2_5_6 __lowercase = audio[:1_0] __lowercase = np.array( [-0.0_0_5_1, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_4, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_2] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowercase = self.get_dummy_components() __lowercase = PNDMScheduler(skip_prk_steps=lowercase__ ) __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = '''A hammer hitting a wooden surface''' # test num_waveforms_per_prompt=1 (default) __lowercase = audioldm_pipe(lowercase__ ,num_inference_steps=2 ).audios assert audios.shape == (1, 2_5_6) # test num_waveforms_per_prompt=1 (default) for batch of prompts __lowercase = 2 __lowercase = audioldm_pipe([prompt] * batch_size ,num_inference_steps=2 ).audios assert audios.shape == (batch_size, 2_5_6) # test num_waveforms_per_prompt for single prompt __lowercase = 2 __lowercase = audioldm_pipe(lowercase__ ,num_inference_steps=2 ,num_waveforms_per_prompt=lowercase__ ).audios assert audios.shape == (num_waveforms_per_prompt, 2_5_6) # test num_waveforms_per_prompt for batch of prompts __lowercase = 2 __lowercase = audioldm_pipe( [prompt] * batch_size ,num_inference_steps=2 ,num_waveforms_per_prompt=lowercase__ ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 2_5_6) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowercase = self.get_dummy_components() __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = audioldm_pipe.vocoder.config.sampling_rate __lowercase = self.get_dummy_inputs(lowercase__ ) __lowercase = audioldm_pipe(audio_length_in_s=0.0_1_6 ,**lowercase__ ) __lowercase = output.audios[0] assert audio.ndim == 1 assert len(lowercase__ ) / vocoder_sampling_rate == 0.0_1_6 __lowercase = audioldm_pipe(audio_length_in_s=0.0_3_2 ,**lowercase__ ) __lowercase = output.audios[0] assert audio.ndim == 1 assert len(lowercase__ ) / vocoder_sampling_rate == 0.0_3_2 def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.get_dummy_components() __lowercase = AudioLDMPipeline(**lowercase__ ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = ['''hey'''] __lowercase = audioldm_pipe(lowercase__ ,num_inference_steps=1 ) __lowercase = output.audios.shape assert audio_shape == (1, 2_5_6) __lowercase = audioldm_pipe.vocoder.config config.model_in_dim *= 2 __lowercase = SpeechTaHifiGan(lowercase__ ).to(lowercase__ ) __lowercase = audioldm_pipe(lowercase__ ,num_inference_steps=1 ) __lowercase = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 2_5_6) def SCREAMING_SNAKE_CASE ( self : List[Any] ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self._test_inference_batch_single_identical(test_mean_pixel_difference=lowercase__ ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() ,reason='''XFormers attention is only available with CUDA and `xformers` installed''' ,) def SCREAMING_SNAKE_CASE ( self : List[str] ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowercase__ ) @slow class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Dict ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : str ,lowercase__ : Dict="cpu" ,lowercase__ : List[Any]=torch.floataa ,lowercase__ : Tuple=0 ): __lowercase = torch.Generator(device=lowercase__ ).manual_seed(lowercase__ ) __lowercase = np.random.RandomState(lowercase__ ).standard_normal((1, 8, 1_2_8, 1_6) ) __lowercase = torch.from_numpy(lowercase__ ).to(device=lowercase__ ,dtype=lowercase__ ) __lowercase = { '''prompt''': '''A hammer hitting a wooden surface''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 2.5, } return inputs def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = AudioLDMPipeline.from_pretrained('''cvssp/audioldm''' ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_inputs(lowercase__ ) __lowercase = 2_5 __lowercase = audioldm_pipe(**lowercase__ ).audios[0] assert audio.ndim == 1 assert len(lowercase__ ) == 8_1_9_2_0 __lowercase = audio[7_7_2_3_0:7_7_2_4_0] __lowercase = np.array( [-0.4_8_8_4, -0.4_6_0_7, 0.0_0_2_3, 0.5_0_0_7, 0.5_8_9_6, 0.5_1_5_1, 0.3_8_1_3, -0.0_2_0_8, -0.3_6_8_7, -0.4_3_1_5] ) __lowercase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = AudioLDMPipeline.from_pretrained('''cvssp/audioldm''' ) __lowercase = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) __lowercase = audioldm_pipe.to(lowercase__ ) audioldm_pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = self.get_inputs(lowercase__ ) __lowercase = audioldm_pipe(**lowercase__ ).audios[0] assert audio.ndim == 1 assert len(lowercase__ ) == 8_1_9_2_0 __lowercase = audio[2_7_7_8_0:2_7_7_9_0] __lowercase = np.array([-0.2_1_3_1, -0.0_8_7_3, -0.0_1_2_4, -0.0_1_8_9, 0.0_5_6_9, 0.1_3_7_3, 0.1_8_8_3, 0.2_8_8_6, 0.3_2_9_7, 0.2_2_1_2] ) __lowercase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
52
'''simple docstring''' import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = OpenAIGPTTokenizer SCREAMING_SNAKE_CASE : str = OpenAIGPTTokenizerFast SCREAMING_SNAKE_CASE : Optional[int] = True SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : int ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] __lowercase = dict(zip(lowercase__ ,range(len(lowercase__ ) ) ) ) __lowercase = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ) as fp: fp.write(json.dumps(lowercase__ ) ) with open(self.merges_file ,'''w''' ) as fp: fp.write('''\n'''.join(lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[Any] ): return "lower newer", "lower newer" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = OpenAIGPTTokenizer(self.vocab_file ,self.merges_file ) __lowercase = '''lower''' __lowercase = ['''low''', '''er</w>'''] __lowercase = tokenizer.tokenize(lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) __lowercase = tokens + ['''<unk>'''] __lowercase = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase__ ) ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Dict=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(lowercase__ ,**lowercase__ ) # Simple input __lowercase = '''This is a simple input''' __lowercase = ['''This is a simple input 1''', '''This is a simple input 2'''] __lowercase = ('''This is a simple input''', '''This is a pair''') __lowercase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(lowercase__ ,tokenizer_r.encode ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Simple input self.assertRaises(lowercase__ ,tokenizer_r.encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Simple input self.assertRaises( lowercase__ ,tokenizer_r.batch_encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ,) # Pair input self.assertRaises(lowercase__ ,tokenizer_r.encode ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Pair input self.assertRaises(lowercase__ ,tokenizer_r.encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Pair input self.assertRaises( lowercase__ ,tokenizer_r.batch_encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass @require_ftfy @require_spacy @require_tokenizers class lowercase_ (lowerCamelCase__ ): """simple docstring""" pass
52
1
'''simple docstring''' import math def __snake_case ( UpperCAmelCase_ : float , UpperCAmelCase_ : float ): return math.pow(UpperCAmelCase_ , 2 ) - a def __snake_case ( UpperCAmelCase_ : float ): return 2 * x def __snake_case ( UpperCAmelCase_ : float ): lowerCamelCase_ = 2.0 while start <= a: lowerCamelCase_ = math.pow(UpperCAmelCase_ , 2 ) return start def __snake_case ( UpperCAmelCase_ : float , UpperCAmelCase_ : int = 9999 , UpperCAmelCase_ : float = 0.00_0000_0000_0001 ): if a < 0: raise ValueError("math domain error" ) lowerCamelCase_ = get_initial_point(UpperCAmelCase_ ) for _ in range(UpperCAmelCase_ ): lowerCamelCase_ = value lowerCamelCase_ = value - fx(UpperCAmelCase_ , UpperCAmelCase_ ) / fx_derivative(UpperCAmelCase_ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
55
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" snake_case_ = '''EncodecFeatureExtractor''' snake_case_ = ('''T5Tokenizer''', '''T5TokenizerFast''') def __init__( self , lowerCamelCase__ , lowerCamelCase__ ) -> int: '''simple docstring''' super().__init__(lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase = self.feature_extractor __lowerCamelCase = False def lowercase_ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True ) -> List[Any]: '''simple docstring''' return self.tokenizer.get_decoder_prompt_ids(task=lowerCamelCase__ , language=lowerCamelCase__ , no_timestamps=lowerCamelCase__ ) def __call__( self , *lowerCamelCase__ , **lowerCamelCase__ ) -> Dict: '''simple docstring''' # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowerCamelCase__ , **lowerCamelCase__ ) __lowerCamelCase = kwargs.pop('audio' , lowerCamelCase__ ) __lowerCamelCase = kwargs.pop('sampling_rate' , lowerCamelCase__ ) __lowerCamelCase = kwargs.pop('text' , lowerCamelCase__ ) if len(lowerCamelCase__ ) > 0: __lowerCamelCase = args[0] __lowerCamelCase = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if text is not None: __lowerCamelCase = self.tokenizer(lowerCamelCase__ , **lowerCamelCase__ ) if audio is not None: __lowerCamelCase = self.feature_extractor(lowerCamelCase__ , *lowerCamelCase__ , sampling_rate=lowerCamelCase__ , **lowerCamelCase__ ) if audio is None: return inputs elif text is None: return audio_inputs else: __lowerCamelCase = audio_inputs['input_values'] if "padding_mask" in audio_inputs: __lowerCamelCase = audio_inputs['padding_mask'] return inputs def lowercase_ ( self , *lowerCamelCase__ , **lowerCamelCase__ ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = kwargs.pop('audio' , lowerCamelCase__ ) __lowerCamelCase = kwargs.pop('padding_mask' , lowerCamelCase__ ) if len(lowerCamelCase__ ) > 0: __lowerCamelCase = args[0] __lowerCamelCase = args[1:] if audio_values is not None: return self._decode_audio(lowerCamelCase__ , padding_mask=lowerCamelCase__ ) else: return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def lowercase_ ( self , *lowerCamelCase__ , **lowerCamelCase__ ) -> List[Any]: '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> List[np.ndarray]: '''simple docstring''' __lowerCamelCase = to_numpy(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = audio_values.shape if padding_mask is None: return list(lowerCamelCase__ ) __lowerCamelCase = to_numpy(lowerCamelCase__ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __lowerCamelCase = seq_len - padding_mask.shape[-1] __lowerCamelCase = 1 - self.feature_extractor.padding_value __lowerCamelCase = np.pad(lowerCamelCase__ , ((0, 0), (0, difference)) , 'constant' , constant_values=lowerCamelCase__ ) __lowerCamelCase = audio_values.tolist() for i in range(lowerCamelCase__ ): __lowerCamelCase = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __lowerCamelCase = sliced_audio.reshape(lowerCamelCase__ , -1 ) return audio_values
90
0
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __a ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=4 , ) -> Dict: '''simple docstring''' lowercase__: Union[str, Any] = parent lowercase__: List[str] = batch_size lowercase__: List[Any] = seq_length lowercase__: Any = is_training lowercase__: Any = use_attention_mask lowercase__: Any = use_token_type_ids lowercase__: List[Any] = use_labels lowercase__: int = vocab_size lowercase__: int = hidden_size lowercase__: List[str] = num_hidden_layers lowercase__: Dict = num_attention_heads lowercase__: int = intermediate_size lowercase__: List[Any] = hidden_act lowercase__: int = hidden_dropout_prob lowercase__: Tuple = attention_probs_dropout_prob lowercase__: List[Any] = max_position_embeddings lowercase__: Tuple = type_vocab_size lowercase__: Optional[int] = type_sequence_label_size lowercase__: Any = initializer_range lowercase__: int = num_choices def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' lowercase__: Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__: Optional[Any] = None if self.use_attention_mask: lowercase__: Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__: List[Any] = None if self.use_token_type_ids: lowercase__: Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__: List[Any] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' lowercase__: Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__: Dict = config_and_inputs lowercase__: Any = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' lowercase__: List[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__: Optional[int] = config_and_inputs lowercase__: Optional[Any] = True lowercase__: Optional[Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowercase__: str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __a ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): __lowercase : int = True __lowercase : List[Any] = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE__ ( self ) -> Tuple: '''simple docstring''' lowercase__: Dict = FlaxRobertaPreLayerNormModelTester(self ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' for model_class_name in self.all_model_classes: lowercase__: Tuple = model_class_name.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=_a ) lowercase__: Optional[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_a ) @require_flax class __a ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: Optional[Any] = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=_a ) lowercase__: Dict = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) lowercase__: Optional[int] = model(_a )[0] lowercase__: Optional[int] = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _a ) # compare the actual values for a slice. lowercase__: int = np.array( [[[4_0.4_8_8_0, 1_8.0_1_9_9, -5.2_3_6_7], [-1.8_8_7_7, -4.0_8_8_5, 1_0.7_0_8_5], [-2.2_6_1_3, -5.6_1_1_0, 7.2_6_6_5]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1E-4 ) ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> List[Any]: '''simple docstring''' lowercase__: str = FlaxRobertaPreLayerNormModel.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=_a ) lowercase__: int = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) lowercase__: Union[str, Any] = model(_a )[0] # compare the actual values for a slice. lowercase__: Optional[int] = np.array( [[[0.0_2_0_8, -0.0_3_5_6, 0.0_2_3_7], [-0.1_5_6_9, -0.0_4_1_1, -0.2_6_2_6], [0.1_8_7_9, 0.0_1_2_5, -0.0_0_8_9]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1E-4 ) )
361
import torch from diffusers import StableDiffusionPipeline __lowerCAmelCase = '''path-to-your-trained-model''' __lowerCAmelCase = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('''cuda''') __lowerCAmelCase = '''A photo of sks dog in a bucket''' __lowerCAmelCase = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('''dog-bucket.png''')
288
0
'''simple docstring''' from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : Any = 0 if start < end: lowerCAmelCase__ : List[Any] = randint(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : int = a[end] lowerCAmelCase__ : Union[str, Any] = a[pivot] lowerCAmelCase__ : List[str] = temp lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = _in_place_partition(UpperCamelCase , UpperCamelCase , UpperCamelCase ) count += _in_place_quick_sort(UpperCamelCase , UpperCamelCase , p - 1 ) count += _in_place_quick_sort(UpperCamelCase , p + 1 , UpperCamelCase ) return count def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : Optional[int] = randint(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : Optional[int] = a[end] lowerCAmelCase__ : Union[str, Any] = a[pivot] lowerCAmelCase__ : List[str] = temp lowerCAmelCase__ : Dict = start - 1 for index in range(UpperCamelCase , UpperCamelCase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value lowerCAmelCase__ : Tuple = new_pivot_index + 1 lowerCAmelCase__ : str = a[new_pivot_index] lowerCAmelCase__ : Dict = a[index] lowerCAmelCase__ : Dict = temp lowerCAmelCase__ : Union[str, Any] = a[new_pivot_index + 1] lowerCAmelCase__ : Any = a[end] lowerCAmelCase__ : Optional[int] = temp return new_pivot_index + 1, count _lowerCAmelCase = TemporaryFile() _lowerCAmelCase = 100 # 1000 elements are to be sorted _lowerCAmelCase , _lowerCAmelCase = 0, 1 # mean and standard deviation _lowerCAmelCase = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array _lowerCAmelCase = np.load(outfile) _lowerCAmelCase = len(M) - 1 _lowerCAmelCase = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
37
'''simple docstring''' from pathlib import Path import fire def lowercase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : str , lowerCAmelCase__ : int ): """simple docstring""" __UpperCAmelCase : List[str] = Path(lowerCAmelCase__ ) __UpperCAmelCase : str = Path(lowerCAmelCase__ ) dest_dir.mkdir(exist_ok=lowerCAmelCase__ ) for path in src_dir.iterdir(): __UpperCAmelCase : str = [x.rstrip() for x in list(path.open().readlines() )][:n] __UpperCAmelCase : Optional[int] = dest_dir.joinpath(path.name ) print(lowerCAmelCase__ ) dest_path.open("""w""" ).write("""\n""".join(lowerCAmelCase__ ) ) if __name__ == "__main__": fire.Fire(minify)
254
0
import argparse import os import re _UpperCamelCase = '''src/diffusers''' # Pattern that looks at the indentation in a line. _UpperCamelCase = re.compile(R'''^(\s*)\S''') # Pattern that matches `"key":" and puts `key` in group 0. _UpperCamelCase = re.compile(R'''^\s*"([^"]+)":''') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _UpperCamelCase = re.compile(R'''^\s*_import_structure\["([^"]+)"\]''') # Pattern that matches `"key",` and puts `key` in group 0. _UpperCamelCase = re.compile(R'''^\s*"([^"]+)",\s*$''') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _UpperCamelCase = re.compile(R'''\[([^\]]+)\]''') def UpperCamelCase_( snake_case__: Tuple ) -> List[str]: UpperCAmelCase__ = _re_indent.search(snake_case__ ) return "" if search is None else search.groups()[0] def UpperCamelCase_( snake_case__: Any , snake_case__: int="" , snake_case__: int=None , snake_case__: Any=None ) -> List[Any]: UpperCAmelCase__ = 0 UpperCAmelCase__ = code.split('\n' ) if start_prompt is not None: while not lines[index].startswith(snake_case__ ): index += 1 UpperCAmelCase__ = ['\n'.join(lines[:index] )] else: UpperCAmelCase__ = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). UpperCAmelCase__ = [lines[index]] index += 1 while index < len(snake_case__ ) and (end_prompt is None or not lines[index].startswith(snake_case__ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(snake_case__ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ' ' ): current_block.append(lines[index] ) blocks.append('\n'.join(snake_case__ ) ) if index < len(snake_case__ ) - 1: UpperCAmelCase__ = [lines[index + 1]] index += 1 else: UpperCAmelCase__ = [] else: blocks.append('\n'.join(snake_case__ ) ) UpperCAmelCase__ = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(snake_case__ ) > 0: blocks.append('\n'.join(snake_case__ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(snake_case__ ): blocks.append('\n'.join(lines[index:] ) ) return blocks def UpperCamelCase_( snake_case__: Optional[int] ) -> Dict: def _inner(snake_case__: str ): return key(snake_case__ ).lower().replace('_' , '' ) return _inner def UpperCamelCase_( snake_case__: int , snake_case__: List[Any]=None ) -> int: # If no key is provided, we use a noop. def noop(snake_case__: Union[str, Any] ): return x if key is None: UpperCAmelCase__ = noop # Constants are all uppercase, they go first. UpperCAmelCase__ = [obj for obj in objects if key(snake_case__ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. UpperCAmelCase__ = [obj for obj in objects if key(snake_case__ )[0].isupper() and not key(snake_case__ ).isupper()] # Functions begin with a lowercase, they go last. UpperCAmelCase__ = [obj for obj in objects if not key(snake_case__ )[0].isupper()] UpperCAmelCase__ = ignore_underscore(snake_case__ ) return sorted(snake_case__ , key=snake_case__ ) + sorted(snake_case__ , key=snake_case__ ) + sorted(snake_case__ , key=snake_case__ ) def UpperCamelCase_( snake_case__: Optional[Any] ) -> int: # This inner function sort imports between [ ]. def _replace(snake_case__: Dict ): UpperCAmelCase__ = match.groups()[0] if "," not in imports: return f"[{imports}]" UpperCAmelCase__ = [part.strip().replace('"' , '' ) for part in imports.split(',' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase__ = keys[:-1] return "[" + ", ".join([f"\"{k}\"" for k in sort_objects(snake_case__ )] ) + "]" UpperCAmelCase__ = import_statement.split('\n' ) if len(snake_case__ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. UpperCAmelCase__ = 2 if lines[1].strip() == '[' else 1 UpperCAmelCase__ = [(i, _re_strip_line.search(snake_case__ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] UpperCAmelCase__ = sort_objects(snake_case__ , key=lambda snake_case__ : x[1] ) UpperCAmelCase__ = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(snake_case__ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: UpperCAmelCase__ = _re_bracket_content.sub(_replace , lines[1] ) else: UpperCAmelCase__ = [part.strip().replace('"' , '' ) for part in lines[1].split(',' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase__ = keys[:-1] UpperCAmelCase__ = get_indent(lines[1] ) + ', '.join([f"\"{k}\"" for k in sort_objects(snake_case__ )] ) return "\n".join(snake_case__ ) else: # Finally we have to deal with imports fitting on one line UpperCAmelCase__ = _re_bracket_content.sub(_replace , snake_case__ ) return import_statement def UpperCamelCase_( snake_case__: Union[str, Any] , snake_case__: Union[str, Any]=True ) -> Optional[int]: with open(snake_case__ , 'r' ) as f: UpperCAmelCase__ = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 UpperCAmelCase__ = split_code_in_indented_blocks( snake_case__ , start_prompt='_import_structure = {' , end_prompt='if TYPE_CHECKING:' ) # We ignore block 0 (everything until start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(snake_case__ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. UpperCAmelCase__ = main_blocks[block_idx] UpperCAmelCase__ = block.split('\n' ) # Get to the start of the imports. UpperCAmelCase__ = 0 while line_idx < len(snake_case__ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: UpperCAmelCase__ = len(snake_case__ ) else: line_idx += 1 if line_idx >= len(snake_case__ ): continue # Ignore beginning and last line: they don't contain anything. UpperCAmelCase__ = '\n'.join(block_lines[line_idx:-1] ) UpperCAmelCase__ = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. UpperCAmelCase__ = split_code_in_indented_blocks(snake_case__ , indent_level=snake_case__ ) # We have two categories of import key: list or _import_structure[key].append/extend UpperCAmelCase__ = _re_direct_key if '_import_structure' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. UpperCAmelCase__ = [(pattern.search(snake_case__ ).groups()[0] if pattern.search(snake_case__ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. UpperCAmelCase__ = [(i, key) for i, key in enumerate(snake_case__ ) if key is not None] UpperCAmelCase__ = [x[0] for x in sorted(snake_case__ , key=lambda snake_case__ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. UpperCAmelCase__ = 0 UpperCAmelCase__ = [] for i in range(len(snake_case__ ) ): if keys[i] is None: reordered_blocks.append(internal_blocks[i] ) else: UpperCAmelCase__ = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reordered_blocks.append(snake_case__ ) count += 1 # And we put our main block back together with its first and last line. UpperCAmelCase__ = '\n'.join(block_lines[:line_idx] + reordered_blocks + [block_lines[-1]] ) if code != "\n".join(snake_case__ ): if check_only: return True else: print(f"Overwriting {file}." ) with open(snake_case__ , 'w' ) as f: f.write('\n'.join(snake_case__ ) ) def UpperCamelCase_( snake_case__: Optional[Any]=True ) -> Tuple: UpperCAmelCase__ = [] for root, _, files in os.walk(snake_case__ ): if "__init__.py" in files: UpperCAmelCase__ = sort_imports(os.path.join(snake_case__ , '__init__.py' ) , check_only=snake_case__ ) if result: UpperCAmelCase__ = [os.path.join(snake_case__ , '__init__.py' )] if len(snake_case__ ) > 0: raise ValueError(f"Would overwrite {len(snake_case__ )} files, run `make style`." ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') _UpperCamelCase = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
335
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class lowercase ( unittest.TestCase ): '''simple docstring''' def __init__(self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=16 , __a=2 , __a=0.02 , __a=4 , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_choices def UpperCamelCase__ (self ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_attention_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = True UpperCAmelCase__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowercase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ = FlaxRobertaModelTester(self ) @slow def UpperCamelCase__ (self ) -> str: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase__ = model_class_name.from_pretrained('roberta-base' , from_pt=__a ) UpperCAmelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__a )
335
1
'''simple docstring''' import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __snake_case = get_tests_dir('''fixtures/test_sentencepiece_no_bos.model''') @require_sentencepiece @require_tokenizers class lowercase ( __snake_case , unittest.TestCase ): """simple docstring""" _a = PegasusTokenizer _a = PegasusTokenizerFast _a = True _a = True def lowerCAmelCase__ ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase__ :str = PegasusTokenizer(a_ ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCAmelCase__ ( self ): '''simple docstring''' return PegasusTokenizer.from_pretrained('''google/pegasus-large''' ) def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' return PegasusTokenizer.from_pretrained(self.tmpdirname , **a_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return ("This is a test", "This is a test") def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = '''</s>''' UpperCamelCase__ :str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''</s>''' ) self.assertEqual(vocab_keys[-1] , '''v''' ) self.assertEqual(len(a_ ) , 1103 ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1103 ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCamelCase__ :List[str] = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCamelCase__ :Optional[Any] = ( '''Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important''' ''' </s> <pad> <pad> <pad>''' ) UpperCamelCase__ :Optional[Any] = rust_tokenizer([raw_input_str] , return_tensors=a_ , add_special_tokens=a_ ).input_ids[0] UpperCamelCase__ :Dict = py_tokenizer([raw_input_str] , return_tensors=a_ , add_special_tokens=a_ ).input_ids[0] self.assertListEqual(a_ , a_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word UpperCamelCase__ :Optional[Any] = '''<mask_1> To ensure a <mask_2> flow of bank resolutions.''' UpperCamelCase__ :str = [2, 413, 615, 114, 3, 1971, 113, 1679, 10710, 107, 1] UpperCamelCase__ :Optional[Any] = tokenizer([raw_input_str] , return_tensors=a_ ).input_ids[0] self.assertListEqual(a_ , a_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[Any] = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 96103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1024 UpperCamelCase__ :Optional[int] = '''To ensure a smooth flow of bank resolutions.''' UpperCamelCase__ :List[str] = [413, 615, 114, 2291, 1971, 113, 1679, 10710, 107, 1] UpperCamelCase__ :Union[str, Any] = tokenizer([raw_input_str] , return_tensors=a_ ).input_ids[0] self.assertListEqual(a_ , a_ ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[Any] = ['''This is going to be way too long.''' * 150, '''short example'''] UpperCamelCase__ :Optional[Any] = ['''not super long but more than 5 tokens''', '''tiny'''] UpperCamelCase__ :List[Any] = self._large_tokenizer(a_ , padding=a_ , truncation=a_ , return_tensors='''pt''' ) UpperCamelCase__ :Dict = self._large_tokenizer( text_target=a_ , max_length=5 , padding=a_ , truncation=a_ , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 1024) assert batch.attention_mask.shape == (2, 1024) assert targets["input_ids"].shape == (2, 5) assert len(a_ ) == 2 # input_ids, attention_mask. @slow def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = {'''input_ids''': [[38979, 143, 18485, 606, 130, 26669, 87686, 121, 54189, 1129, 111, 26669, 87686, 121, 9114, 14787, 121, 13249, 158, 592, 956, 121, 14621, 31576, 143, 62613, 108, 9688, 930, 43430, 11562, 62613, 304, 108, 11443, 897, 108, 9314, 17415, 63399, 108, 11443, 7614, 18316, 118, 4284, 7148, 12430, 143, 1400, 25703, 158, 111, 4284, 7148, 11772, 143, 21297, 1064, 158, 122, 204, 3506, 1754, 1133, 14787, 1581, 115, 33224, 4482, 111, 1355, 110, 29173, 317, 50833, 108, 20147, 94665, 111, 77198, 107, 1], [110, 62613, 117, 638, 112, 1133, 121, 20098, 1355, 79050, 13872, 135, 1596, 53541, 1352, 141, 13039, 5542, 124, 302, 518, 111, 268, 2956, 115, 149, 4427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1235, 2799, 18289, 17780, 204, 109, 9474, 1296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name='''google/bigbird-pegasus-large-arxiv''' , revision='''ba85d0851d708441f91440d509690f1ab6353415''' , ) @require_sentencepiece @require_tokenizers class lowercase ( __snake_case , unittest.TestCase ): """simple docstring""" _a = PegasusTokenizer _a = PegasusTokenizerFast _a = True _a = True def lowerCAmelCase__ ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase__ :Dict = PegasusTokenizer(a_ , offset=0 , mask_token_sent=a_ , mask_token='''[MASK]''' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCAmelCase__ ( self ): '''simple docstring''' return PegasusTokenizer.from_pretrained('''google/bigbird-pegasus-large-arxiv''' ) def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' return PegasusTokenizer.from_pretrained(self.tmpdirname , **a_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return ("This is a test", "This is a test") def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCamelCase__ :List[str] = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCamelCase__ :Union[str, Any] = ( '''Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>''' ''' <pad> <pad> <pad>''' ) UpperCamelCase__ :List[str] = rust_tokenizer([raw_input_str] , return_tensors=a_ , add_special_tokens=a_ ).input_ids[0] UpperCamelCase__ :List[str] = py_tokenizer([raw_input_str] , return_tensors=a_ , add_special_tokens=a_ ).input_ids[0] self.assertListEqual(a_ , a_ ) @require_torch def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = ['''This is going to be way too long.''' * 1000, '''short example'''] UpperCamelCase__ :Optional[Any] = ['''not super long but more than 5 tokens''', '''tiny'''] UpperCamelCase__ :str = self._large_tokenizer(a_ , padding=a_ , truncation=a_ , return_tensors='''pt''' ) UpperCamelCase__ :List[Any] = self._large_tokenizer( text_target=a_ , max_length=5 , padding=a_ , truncation=a_ , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 4096) assert batch.attention_mask.shape == (2, 4096) assert targets["input_ids"].shape == (2, 5) assert len(a_ ) == 2 # input_ids, attention_mask. def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = ( '''This is an example string that is used to test the original TF implementation against the HF''' ''' implementation''' ) UpperCamelCase__ :Optional[Any] = self._large_tokenizer(a_ ).input_ids self.assertListEqual( a_ , [182, 117, 142, 587, 4211, 120, 117, 263, 112, 804, 109, 856, 25016, 3137, 464, 109, 26955, 3137, 1] , )
97
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = StableDiffusionKDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' ) __snake_case : List[str] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler('''sample_euler''' ) __snake_case : Tuple = '''A painting of a squirrel eating a burger''' __snake_case : str = torch.manual_seed(0 ) __snake_case : int = sd_pipe([prompt] , generator=a_ , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''' ) __snake_case : List[str] = output.images __snake_case : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) __snake_case : Dict = np.array([0.0447, 0.0492, 0.0468, 0.0408, 0.0383, 0.0408, 0.0354, 0.0380, 0.0339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) __snake_case : Union[str, Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler('''sample_euler''' ) __snake_case : str = '''A painting of a squirrel eating a burger''' __snake_case : List[str] = torch.manual_seed(0 ) __snake_case : Dict = sd_pipe([prompt] , generator=a_ , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''' ) __snake_case : Any = output.images __snake_case : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) __snake_case : Tuple = np.array([0.1237, 0.1320, 0.1438, 0.1359, 0.1390, 0.1132, 0.1277, 0.1175, 0.1112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) __snake_case : Union[str, Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler('''sample_dpmpp_2m''' ) __snake_case : int = '''A painting of a squirrel eating a burger''' __snake_case : Optional[Any] = torch.manual_seed(0 ) __snake_case : Dict = sd_pipe( [prompt] , generator=a_ , guidance_scale=7.5 , num_inference_steps=15 , output_type='''np''' , use_karras_sigmas=a_ , ) __snake_case : Optional[int] = output.images __snake_case : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) __snake_case : Any = np.array( [0.1138_1689, 0.1211_2921, 0.138_9457, 0.1254_9606, 0.124_4964, 0.1083_1517, 0.1156_2866, 0.1086_7816, 0.1049_9048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
102
0
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version UpperCAmelCase__ = version.parse(importlib_metadata.version("nltk")) if NLTK_VERSION >= version.Version("3.6.4"): from nltk import word_tokenize UpperCAmelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' UpperCAmelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' UpperCAmelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): def _lowerCamelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence'), 'references': datasets.Value('string' , id='sequence'), }) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def _lowerCamelCase ( self : List[str] , A : Optional[int]) -> Union[str, Any]: """simple docstring""" import nltk nltk.download('wordnet') if NLTK_VERSION >= version.Version('3.6.5'): nltk.download('punkt') if NLTK_VERSION >= version.Version('3.6.6'): nltk.download('omw-1.4') def _lowerCamelCase ( self : Any , A : List[Any] , A : List[str] , A : List[str]=0.9 , A : Optional[int]=3 , A : Union[str, Any]=0.5) -> str: """simple docstring""" if NLTK_VERSION >= version.Version('3.6.5'): _UpperCAmelCase = [ meteor_score.single_meteor_score( word_tokenize(lowercase_) , word_tokenize(lowercase_) , alpha=lowercase_ , beta=lowercase_ , gamma=lowercase_) for ref, pred in zip(lowercase_ , lowercase_) ] else: _UpperCAmelCase = [ meteor_score.single_meteor_score(lowercase_ , lowercase_ , alpha=lowercase_ , beta=lowercase_ , gamma=lowercase_) for ref, pred in zip(lowercase_ , lowercase_) ] return {"meteor": np.mean(lowercase_)}
356
def A ( _UpperCAmelCase : int ) -> "list[int]": '''simple docstring''' if upper_limit < 0: raise ValueError('Limit for the Catalan sequence must be ≥ 0' ) _UpperCAmelCase = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 _UpperCAmelCase = 1 if upper_limit > 0: _UpperCAmelCase = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(_UpperCAmelCase ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("\n********* Catalan Numbers Using Dynamic Programming ************\n") print("\n*** Enter -1 at any time to quit ***") print("\nEnter the upper limit (≥ 0) for the Catalan number sequence: ", end="") try: while True: UpperCAmelCase__ = int(input().strip()) if N < 0: print("\n********* Goodbye!! ************") break else: print(f"""The Catalan numbers from 0 through {N} are:""") print(catalan_numbers(N)) print("Try another upper limit for the sequence: ", end="") except (NameError, ValueError): print("\n********* Invalid input, goodbye! ************\n") import doctest doctest.testmod()
290
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase = logging.get_logger(__name__) __lowerCamelCase = { """google/canine-s""": """https://huggingface.co/google/canine-s/resolve/main/config.json""", # See all CANINE models at https://huggingface.co/models?filter=canine } class UpperCAmelCase ( A_ ): A__ : List[Any] = "canine" def __init__(self : Dict , snake_case__ : Dict=7_68 , snake_case__ : Tuple=12 , snake_case__ : Optional[int]=12 , snake_case__ : Optional[Any]=30_72 , snake_case__ : List[Any]="gelu" , snake_case__ : Dict=0.1 , snake_case__ : Optional[Any]=0.1 , snake_case__ : List[str]=1_63_84 , snake_case__ : List[Any]=16 , snake_case__ : List[Any]=0.02 , snake_case__ : Tuple=1e-12 , snake_case__ : Tuple=0 , snake_case__ : Optional[int]=0XE_0_0_0 , snake_case__ : Dict=0XE_0_0_1 , snake_case__ : int=4 , snake_case__ : Union[str, Any]=4 , snake_case__ : Union[str, Any]=8 , snake_case__ : List[str]=1_63_84 , snake_case__ : List[str]=1_28 , **snake_case__ : Optional[Any] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) snake_case : Any = max_position_embeddings snake_case : List[Any] = hidden_size snake_case : Optional[int] = num_hidden_layers snake_case : Tuple = num_attention_heads snake_case : Dict = intermediate_size snake_case : List[str] = hidden_act snake_case : Tuple = hidden_dropout_prob snake_case : List[str] = attention_probs_dropout_prob snake_case : str = initializer_range snake_case : int = type_vocab_size snake_case : List[Any] = layer_norm_eps # Character config: snake_case : str = downsampling_rate snake_case : Dict = upsampling_kernel_size snake_case : List[str] = num_hash_functions snake_case : Optional[int] = num_hash_buckets snake_case : Dict = local_transformer_stride
59
import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class UpperCAmelCase ( unittest.TestCase ): def _SCREAMING_SNAKE_CASE (self : Any ) -> List[str]: '''simple docstring''' snake_case : int = tempfile.mkdtemp() # fmt: off snake_case : Optional[int] = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest"] # fmt: on snake_case : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) snake_case : int = { "do_resize": True, "size": {"height": 18, "width": 18}, "do_normalize": True, "image_mean": [0.5, 0.5, 0.5], "image_std": [0.5, 0.5, 0.5], } snake_case : Optional[Any] = os.path.join(self.tmpdirname , snake_case__ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(snake_case__ , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] , **snake_case__ : str ) -> Optional[int]: '''simple docstring''' return BertTokenizer.from_pretrained(self.tmpdirname , **snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] , **snake_case__ : List[str] ) -> int: '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> Dict: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _SCREAMING_SNAKE_CASE (self : Dict ) -> str: '''simple docstring''' snake_case : List[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] snake_case : Optional[int] = [Image.fromarray(np.moveaxis(snake_case__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : Dict = self.get_tokenizer() snake_case : Optional[Any] = self.get_image_processor() snake_case : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) processor.save_pretrained(self.tmpdirname ) snake_case : Any = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Any ) -> Optional[Any]: '''simple docstring''' snake_case : str = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) snake_case : Optional[int] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) snake_case : Tuple = self.get_image_processor(do_normalize=snake_case__ , padding_value=1.0 ) snake_case : List[str] = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=snake_case__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> int: '''simple docstring''' snake_case : str = self.get_image_processor() snake_case : Optional[int] = self.get_tokenizer() snake_case : List[Any] = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) snake_case : Optional[Any] = self.prepare_image_inputs() snake_case : str = image_processor(snake_case__ , return_tensors="np" ) snake_case : Any = processor(images=snake_case__ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _SCREAMING_SNAKE_CASE (self : Tuple ) -> Optional[Any]: '''simple docstring''' snake_case : Dict = self.get_image_processor() snake_case : int = self.get_tokenizer() snake_case : Any = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) snake_case : Tuple = "lower newer" snake_case : Tuple = processor(text=snake_case__ ) snake_case : Union[str, Any] = tokenizer(snake_case__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _SCREAMING_SNAKE_CASE (self : Dict ) -> Optional[int]: '''simple docstring''' snake_case : List[Any] = self.get_image_processor() snake_case : Dict = self.get_tokenizer() snake_case : Dict = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) snake_case : int = "lower newer" snake_case : Dict = self.prepare_image_inputs() snake_case : Union[str, Any] = processor(text=snake_case__ , images=snake_case__ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "token_type_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with self.assertRaises(snake_case__ ): processor() def _SCREAMING_SNAKE_CASE (self : str ) -> Tuple: '''simple docstring''' snake_case : Tuple = self.get_image_processor() snake_case : Optional[Any] = self.get_tokenizer() snake_case : Tuple = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) snake_case : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] snake_case : List[Any] = processor.batch_decode(snake_case__ ) snake_case : Union[str, Any] = tokenizer.batch_decode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> List[str]: '''simple docstring''' snake_case : str = self.get_image_processor() snake_case : Union[str, Any] = self.get_tokenizer() snake_case : Any = VisionTextDualEncoderProcessor(tokenizer=snake_case__ , image_processor=snake_case__ ) snake_case : Optional[Any] = "lower newer" snake_case : List[Any] = self.prepare_image_inputs() snake_case : Tuple = processor(text=snake_case__ , images=snake_case__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
59
1
"""simple docstring""" from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def a__ ( SCREAMING_SNAKE_CASE : Dict[str, torch.Tensor] ): '''simple docstring''' lowerCAmelCase : str = [] lowerCAmelCase : str = [] lowerCAmelCase : Optional[int] = [] for rt in rc.restypes: lowerCAmelCase : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) lowerCAmelCase : int = {name: i for i, name in enumerate(_a )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) lowerCAmelCase : Tuple = torch.tensor( _a , dtype=torch.intaa , device=protein["aatype"].device , ) lowerCAmelCase : List[str] = torch.tensor( _a , dtype=torch.intaa , device=protein["aatype"].device , ) lowerCAmelCase : Tuple = torch.tensor( _a , dtype=torch.floataa , device=protein["aatype"].device , ) lowerCAmelCase : Optional[Any] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein lowerCAmelCase : str = restype_atomaa_to_atomaa[protein_aatype] lowerCAmelCase : List[Any] = restype_atomaa_mask[protein_aatype] lowerCAmelCase : Tuple = residx_atomaa_mask lowerCAmelCase : Optional[Any] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask lowerCAmelCase : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["aatype"].device ) for restype, restype_letter in enumerate(rc.restypes ): lowerCAmelCase : int = rc.restype_atoa[restype_letter] lowerCAmelCase : List[str] = rc.residue_atoms[restype_name] for atom_name in atom_names: lowerCAmelCase : Tuple = rc.atom_order[atom_name] lowerCAmelCase : Any = 1 lowerCAmelCase : List[str] = restype_atomaa_mask[protein_aatype] lowerCAmelCase : Dict = residx_atomaa_mask return protein def a__ ( SCREAMING_SNAKE_CASE : Dict[str, torch.Tensor] ): '''simple docstring''' lowerCAmelCase : Optional[int] = tree_map(lambda SCREAMING_SNAKE_CASE : torch.tensor(_a , device=batch["aatype"].device ) , _a , np.ndarray ) lowerCAmelCase : Optional[Any] = tensor_tree_map(lambda SCREAMING_SNAKE_CASE : np.array(_a ) , make_atomaa_masks(_a ) ) return out
365
"""simple docstring""" import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( lowercase , unittest.TestCase ): """simple docstring""" a : List[str] =None a : List[Any] =BloomTokenizerFast a : Optional[int] =BloomTokenizerFast a : Optional[Any] =True a : Dict =False a : Optional[Any] ="tokenizer_file" a : Optional[int] ={"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def lowercase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase : Tuple = BloomTokenizerFast.from_pretrained("bigscience/tokenizer" ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase__ ( self , **snake_case__ ): """simple docstring""" kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **snake_case__ ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = self.get_rust_tokenizer() lowerCAmelCase : List[Any] = ["The quick brown fox</s>", "jumps over the lazy dog</s>"] lowerCAmelCase : str = [[2_175, 23_714, 73_173, 144_252, 2], [77, 132_619, 3_478, 368, 109_586, 35_433, 2]] lowerCAmelCase : Optional[int] = tokenizer.batch_encode_plus(snake_case__ )["input_ids"] self.assertListEqual(snake_case__ , snake_case__ ) lowerCAmelCase : Optional[int] = tokenizer.batch_decode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) def lowercase__ ( self , snake_case__=6 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input lowerCAmelCase : str = "This is a simple input" lowerCAmelCase : Tuple = ["This is a simple input 1", "This is a simple input 2"] lowerCAmelCase : Any = ("This is a simple input", "This is a pair") lowerCAmelCase : Tuple = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests try: tokenizer_r.encode(snake_case__ , max_length=snake_case__ ) tokenizer_r.encode_plus(snake_case__ , max_length=snake_case__ ) tokenizer_r.batch_encode_plus(snake_case__ , max_length=snake_case__ ) tokenizer_r.encode(snake_case__ , max_length=snake_case__ ) tokenizer_r.batch_encode_plus(snake_case__ , max_length=snake_case__ ) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding" ) lowerCAmelCase : Tuple = None # Hotfixing padding = None self.assertRaises(snake_case__ , tokenizer_r.encode , snake_case__ , max_length=snake_case__ , padding="max_length" ) # Simple input self.assertRaises(snake_case__ , tokenizer_r.encode_plus , snake_case__ , max_length=snake_case__ , padding="max_length" ) # Simple input self.assertRaises( snake_case__ , tokenizer_r.batch_encode_plus , snake_case__ , max_length=snake_case__ , padding="max_length" , ) # Pair input self.assertRaises(snake_case__ , tokenizer_r.encode , snake_case__ , max_length=snake_case__ , padding="max_length" ) # Pair input self.assertRaises(snake_case__ , tokenizer_r.encode_plus , snake_case__ , max_length=snake_case__ , padding="max_length" ) # Pair input self.assertRaises( snake_case__ , tokenizer_r.batch_encode_plus , snake_case__ , max_length=snake_case__ , padding="max_length" , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = self.get_rust_tokenizer() lowerCAmelCase : int = load_dataset("xnli" , "all_languages" , split="test" , streaming=snake_case__ ) lowerCAmelCase : Tuple = next(iter(snake_case__ ) )["premise"] # pick up one data lowerCAmelCase : Optional[Any] = list(sample_data.values() ) lowerCAmelCase : int = list(map(tokenizer.encode , snake_case__ ) ) lowerCAmelCase : List[Any] = [tokenizer.decode(snake_case__ , clean_up_tokenization_spaces=snake_case__ ) for x in output_tokens] self.assertListEqual(snake_case__ , snake_case__ ) def lowercase__ ( self ): """simple docstring""" self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
133
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : Any =logging.get_logger(__name__) __lowerCAmelCase : Tuple ={ 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class _lowercase ( A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = '''unispeech-sat''' def __init__( self :int , lowerCAmelCase__ :Any=32 , lowerCAmelCase__ :int=768 , lowerCAmelCase__ :Optional[int]=12 , lowerCAmelCase__ :str=12 , lowerCAmelCase__ :Dict=3_072 , lowerCAmelCase__ :int="gelu" , lowerCAmelCase__ :str=0.1 , lowerCAmelCase__ :Optional[Any]=0.1 , lowerCAmelCase__ :Dict=0.1 , lowerCAmelCase__ :str=0.0 , lowerCAmelCase__ :str=0.0 , lowerCAmelCase__ :Optional[int]=0.1 , lowerCAmelCase__ :List[Any]=0.1 , lowerCAmelCase__ :List[Any]=0.02 , lowerCAmelCase__ :Union[str, Any]=1E-5 , lowerCAmelCase__ :Any="group" , lowerCAmelCase__ :Union[str, Any]="gelu" , lowerCAmelCase__ :List[str]=(512, 512, 512, 512, 512, 512, 512) , lowerCAmelCase__ :Any=(5, 2, 2, 2, 2, 2, 2) , lowerCAmelCase__ :Optional[int]=(10, 3, 3, 3, 3, 2, 2) , lowerCAmelCase__ :str=False , lowerCAmelCase__ :Optional[int]=128 , lowerCAmelCase__ :Optional[int]=16 , lowerCAmelCase__ :str=False , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :str=0.05 , lowerCAmelCase__ :Any=10 , lowerCAmelCase__ :str=2 , lowerCAmelCase__ :Tuple=0.0 , lowerCAmelCase__ :List[Any]=10 , lowerCAmelCase__ :Tuple=0 , lowerCAmelCase__ :List[str]=320 , lowerCAmelCase__ :Optional[Any]=2 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :Optional[int]=100 , lowerCAmelCase__ :Dict=256 , lowerCAmelCase__ :Optional[Any]=256 , lowerCAmelCase__ :Optional[int]=0.1 , lowerCAmelCase__ :Optional[int]="mean" , lowerCAmelCase__ :Any=False , lowerCAmelCase__ :str=False , lowerCAmelCase__ :List[Any]=256 , lowerCAmelCase__ :int=(512, 512, 512, 512, 1_500) , lowerCAmelCase__ :Optional[Any]=(5, 3, 3, 1, 1) , lowerCAmelCase__ :Optional[int]=(1, 2, 3, 1, 1) , lowerCAmelCase__ :Any=512 , lowerCAmelCase__ :List[Any]=0 , lowerCAmelCase__ :List[Any]=1 , lowerCAmelCase__ :List[str]=2 , lowerCAmelCase__ :int=504 , **lowerCAmelCase__ :Any , ) -> Any: super().__init__(**lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_size __SCREAMING_SNAKE_CASE : Any = feat_extract_norm __SCREAMING_SNAKE_CASE : str = feat_extract_activation __SCREAMING_SNAKE_CASE : int = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = conv_bias __SCREAMING_SNAKE_CASE : Union[str, Any] = num_conv_pos_embeddings __SCREAMING_SNAKE_CASE : Any = num_conv_pos_embedding_groups __SCREAMING_SNAKE_CASE : Union[str, Any] = len(self.conv_dim ) __SCREAMING_SNAKE_CASE : Tuple = num_hidden_layers __SCREAMING_SNAKE_CASE : Dict = intermediate_size __SCREAMING_SNAKE_CASE : Optional[int] = hidden_act __SCREAMING_SNAKE_CASE : str = num_attention_heads __SCREAMING_SNAKE_CASE : str = hidden_dropout __SCREAMING_SNAKE_CASE : List[Any] = attention_dropout __SCREAMING_SNAKE_CASE : Optional[int] = activation_dropout __SCREAMING_SNAKE_CASE : Union[str, Any] = feat_proj_dropout __SCREAMING_SNAKE_CASE : Any = final_dropout __SCREAMING_SNAKE_CASE : str = layerdrop __SCREAMING_SNAKE_CASE : Optional[int] = layer_norm_eps __SCREAMING_SNAKE_CASE : List[str] = initializer_range __SCREAMING_SNAKE_CASE : List[Any] = vocab_size __SCREAMING_SNAKE_CASE : int = num_clusters __SCREAMING_SNAKE_CASE : Any = do_stable_layer_norm __SCREAMING_SNAKE_CASE : Tuple = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __SCREAMING_SNAKE_CASE : Tuple = apply_spec_augment __SCREAMING_SNAKE_CASE : Dict = mask_time_prob __SCREAMING_SNAKE_CASE : Tuple = mask_time_length __SCREAMING_SNAKE_CASE : Dict = mask_time_min_masks __SCREAMING_SNAKE_CASE : str = mask_feature_prob __SCREAMING_SNAKE_CASE : Tuple = mask_feature_length __SCREAMING_SNAKE_CASE : Tuple = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __SCREAMING_SNAKE_CASE : str = num_codevectors_per_group __SCREAMING_SNAKE_CASE : Dict = num_codevector_groups __SCREAMING_SNAKE_CASE : str = contrastive_logits_temperature __SCREAMING_SNAKE_CASE : List[str] = feat_quantizer_dropout __SCREAMING_SNAKE_CASE : str = num_negatives __SCREAMING_SNAKE_CASE : Union[str, Any] = codevector_dim __SCREAMING_SNAKE_CASE : Tuple = proj_codevector_dim __SCREAMING_SNAKE_CASE : List[Any] = diversity_loss_weight # ctc loss __SCREAMING_SNAKE_CASE : Tuple = ctc_loss_reduction __SCREAMING_SNAKE_CASE : Union[str, Any] = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. __SCREAMING_SNAKE_CASE : Dict = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __SCREAMING_SNAKE_CASE : Any = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = xvector_output_dim @property def __magic_name__( self :Optional[int] ) -> Optional[Any]: return functools.reduce(operator.mul , self.conv_stride , 1 )
9
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split __lowerCAmelCase : List[Any] =datasets.load_iris() __lowerCAmelCase : Tuple =np.array(data['data']) __lowerCAmelCase : Dict =np.array(data['target']) __lowerCAmelCase : List[str] =data['target_names'] __lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase : str =train_test_split(X, y) def _UpperCamelCase ( lowercase__ , lowercase__ ): return np.linalg.norm(np.array(lowercase__ ) - np.array(lowercase__ ) ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=5 ): __SCREAMING_SNAKE_CASE : Optional[int] = zip(lowercase__ , lowercase__ ) # List of distances of all points from the point to be classified __SCREAMING_SNAKE_CASE : Dict = [] for data_point in data: __SCREAMING_SNAKE_CASE : Tuple = euclidean_distance(data_point[0] , lowercase__ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. __SCREAMING_SNAKE_CASE : int = [i[1] for i in sorted(lowercase__ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified __SCREAMING_SNAKE_CASE : Any = Counter(lowercase__ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
9
1
def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> Optional[int]: if height >= 1: move_tower(height - 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) move_disk(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) move_tower(height - 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> Union[str, Any]: print("""moving disk from""" , SCREAMING_SNAKE_CASE__ , """to""" , SCREAMING_SNAKE_CASE__) def A__ ( ) -> Optional[int]: __snake_case: List[Any] = int(input("""Height of hanoi: """).strip()) move_tower(SCREAMING_SNAKE_CASE__ , """A""" , """B""" , """C""") if __name__ == "__main__": main()
293
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __snake_case ( __lowerCamelCase ): '''simple docstring''' def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: Optional[int] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(A , """neck_hidden_sizes""" ) ) self.parent.assertTrue(hasattr(A , """num_attention_heads""" ) ) class __snake_case : '''simple docstring''' def __init__( self : int , A : str , A : Dict=13 , A : str=32 , A : Any=2 , A : Optional[Any]=3 , A : str=640 , A : Tuple=4 , A : Dict="silu" , A : List[Any]=3 , A : Any=32 , A : Any=0.1 , A : int=0.1 , A : Dict=0.1 , A : Optional[Any]=0.02 , A : List[Any]=True , A : Tuple=True , A : Any=10 , A : Optional[int]=None , ): __snake_case: List[Any] = parent __snake_case: Dict = batch_size __snake_case: int = image_size __snake_case: Tuple = patch_size __snake_case: Tuple = num_channels __snake_case: str = last_hidden_size __snake_case: Dict = num_attention_heads __snake_case: Dict = hidden_act __snake_case: Tuple = conv_kernel_size __snake_case: List[str] = output_stride __snake_case: List[str] = hidden_dropout_prob __snake_case: Optional[Any] = attention_probs_dropout_prob __snake_case: int = classifier_dropout_prob __snake_case: List[Any] = use_labels __snake_case: Union[str, Any] = is_training __snake_case: Union[str, Any] = num_labels __snake_case: str = initializer_range __snake_case: List[Any] = scope def UpperCAmelCase__ ( self : List[Any] ): __snake_case: Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case: Tuple = None __snake_case: Any = None if self.use_labels: __snake_case: Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case: str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case: Any = self.get_config() return config, pixel_values, labels, pixel_labels def UpperCAmelCase__ ( self : int ): return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self : str , A : Optional[Any] , A : Any , A : Any , A : Union[str, Any] ): __snake_case: List[Any] = MobileViTModel(config=A ) model.to(A ) model.eval() __snake_case: int = model(A ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCAmelCase__ ( self : str , A : List[Any] , A : Any , A : Any , A : int ): __snake_case: str = self.num_labels __snake_case: Optional[int] = MobileViTForImageClassification(A ) model.to(A ) model.eval() __snake_case: Union[str, Any] = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Optional[int] , A : str , A : Optional[Any] , A : int , A : str ): __snake_case: List[Any] = self.num_labels __snake_case: Dict = MobileViTForSemanticSegmentation(A ) model.to(A ) model.eval() __snake_case: Union[str, Any] = model(A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case: Tuple = model(A , labels=A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCAmelCase__ ( self : Dict ): __snake_case: Tuple = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case: Any = config_and_inputs __snake_case: Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __snake_case ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False def UpperCAmelCase__ ( self : List[str] ): __snake_case: List[Any] = MobileViTModelTester(self ) __snake_case: str = MobileViTConfigTester(self , config_class=A , has_text_modality=A ) def UpperCAmelCase__ ( self : str ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileViT does not use inputs_embeds""" ) def UpperCAmelCase__ ( self : List[Any] ): pass @unittest.skip(reason="""MobileViT does not support input and output embeddings""" ) def UpperCAmelCase__ ( self : Dict ): pass @unittest.skip(reason="""MobileViT does not output attentions""" ) def UpperCAmelCase__ ( self : Optional[Any] ): pass def UpperCAmelCase__ ( self : str ): __snake_case , __snake_case: Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case: Optional[Any] = model_class(A ) __snake_case: int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case: Optional[int] = [*signature.parameters.keys()] __snake_case: List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , A ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCAmelCase__ ( self : Optional[int] ): pass def UpperCAmelCase__ ( self : Dict ): __snake_case: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def UpperCAmelCase__ ( self : Dict ): def check_hidden_states_output(A : List[Any] , A : int , A : Tuple ): __snake_case: List[str] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): __snake_case: str = model(**self._prepare_for_class(A , A ) ) __snake_case: Optional[int] = outputs.hidden_states __snake_case: Any = 5 self.assertEqual(len(A ) , A ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case: Union[str, Any] = 2 for i in range(len(A ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case: List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case: Optional[Any] = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case: Dict = True check_hidden_states_output(A , A , A ) def UpperCAmelCase__ ( self : int ): __snake_case: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) def UpperCAmelCase__ ( self : Union[str, Any] ): __snake_case: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A ) @slow def UpperCAmelCase__ ( self : Union[str, Any] ): for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case: List[Any] = MobileViTModel.from_pretrained(A ) self.assertIsNotNone(A ) def A__ ( ) -> Optional[int]: __snake_case: Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""") return image @require_torch @require_vision class __snake_case ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCAmelCase__ ( self : Dict ): return MobileViTImageProcessor.from_pretrained("""apple/mobilevit-xx-small""" ) if is_vision_available() else None @slow def UpperCAmelCase__ ( self : List[Any] ): __snake_case: Tuple = MobileViTForImageClassification.from_pretrained("""apple/mobilevit-xx-small""" ).to(A ) __snake_case: str = self.default_image_processor __snake_case: Optional[Any] = prepare_img() __snake_case: List[Any] = image_processor(images=A , return_tensors="""pt""" ).to(A ) # forward pass with torch.no_grad(): __snake_case: Dict = model(**A ) # verify the logits __snake_case: List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , A ) __snake_case: Union[str, Any] = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A , atol=1E-4 ) ) @slow def UpperCAmelCase__ ( self : Tuple ): __snake_case: Tuple = MobileViTForSemanticSegmentation.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) __snake_case: List[str] = model.to(A ) __snake_case: Dict = MobileViTImageProcessor.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) __snake_case: List[Any] = prepare_img() __snake_case: List[str] = image_processor(images=A , return_tensors="""pt""" ).to(A ) # forward pass with torch.no_grad(): __snake_case: List[Any] = model(**A ) __snake_case: Optional[int] = outputs.logits # verify the logits __snake_case: Dict = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , A ) __snake_case: Optional[int] = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=A , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A , atol=1E-4 ) ) @slow def UpperCAmelCase__ ( self : Dict ): __snake_case: int = MobileViTForSemanticSegmentation.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) __snake_case: str = model.to(A ) __snake_case: Optional[Any] = MobileViTImageProcessor.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) __snake_case: List[str] = prepare_img() __snake_case: Optional[int] = image_processor(images=A , return_tensors="""pt""" ).to(A ) # forward pass with torch.no_grad(): __snake_case: Dict = model(**A ) __snake_case: List[Any] = outputs.logits.detach().cpu() __snake_case: List[str] = image_processor.post_process_semantic_segmentation(outputs=A , target_sizes=[(50, 60)] ) __snake_case: str = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , A ) __snake_case: int = image_processor.post_process_semantic_segmentation(outputs=A ) __snake_case: Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , A )
293
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": SCREAMING_SNAKE_CASE : Dict = pd.read_csv("sample_data.csv", header=None) SCREAMING_SNAKE_CASE : Dict = df.shape[:1][0] # If you're using some other dataset input the target column SCREAMING_SNAKE_CASE : Optional[int] = df.iloc[:, 1:2] SCREAMING_SNAKE_CASE : List[Any] = actual_data.values.reshape(len_data, 1) SCREAMING_SNAKE_CASE : List[str] = MinMaxScaler().fit_transform(actual_data) SCREAMING_SNAKE_CASE : int = 10 SCREAMING_SNAKE_CASE : Optional[int] = 5 SCREAMING_SNAKE_CASE : Optional[int] = 20 SCREAMING_SNAKE_CASE : List[Any] = len_data - periods * look_back SCREAMING_SNAKE_CASE : Tuple = actual_data[:division] SCREAMING_SNAKE_CASE : Union[str, Any] = actual_data[division - look_back :] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = [], [] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) SCREAMING_SNAKE_CASE : Union[str, Any] = np.array(train_x) SCREAMING_SNAKE_CASE : Tuple = np.array(test_x) SCREAMING_SNAKE_CASE : Dict = np.array([list(i.ravel()) for i in train_y]) SCREAMING_SNAKE_CASE : Union[str, Any] = np.array([list(i.ravel()) for i in test_y]) SCREAMING_SNAKE_CASE : Tuple = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") SCREAMING_SNAKE_CASE : Any = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) SCREAMING_SNAKE_CASE : Any = model.predict(x_test)
21
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> bool: if len(lowerCamelCase_ ) < 2: raise ValueError('Monogons and Digons are not polygons in the Euclidean space' ) if any(i <= 0 for i in nums ): raise ValueError('All values must be greater than 0' ) _lowercase : Tuple = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
21
1
'''simple docstring''' from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' def _lowercase ( self ): """simple docstring""" return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(_lowercase ) def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = self._create_example_records() _lowerCAmelCase = Dataset.from_list(_lowercase ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(_lowercase ): self.assertDictEqual(_lowercase , example_records[i] ) def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = self._create_example_records() _lowerCAmelCase = Dataset.from_list(_lowercase ) _lowerCAmelCase = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def _lowercase ( self ): # checks what happens with missing columns """simple docstring""" _lowerCAmelCase = [{"""col_1""": 1}, {"""col_2""": """x"""}] _lowerCAmelCase = Dataset.from_list(_lowercase ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def _lowercase ( self ): # checks if the type can be inferred from the second record """simple docstring""" _lowerCAmelCase = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _lowerCAmelCase = Dataset.from_list(_lowercase ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = Dataset.from_list([] ) self.assertEqual(len(_lowercase ) , 0 ) self.assertListEqual(dset.column_names , [] )
229
'''simple docstring''' from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): '''simple docstring''' @register_to_config def __init__( self , _lowercase , _lowercase = None , _lowercase = None ): """simple docstring""" super().__init__() _lowerCAmelCase = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" _lowerCAmelCase = torch.zeros(_lowercase , _lowercase ) else: _lowerCAmelCase = None _lowerCAmelCase = torch.nn.Parameter(_lowercase ) class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : VQModel _lowercase : CLIPTextModel _lowercase : CLIPTokenizer _lowercase : TransformeraDModel _lowercase : LearnedClassifierFreeSamplingEmbeddings _lowercase : VQDiffusionScheduler def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=_lowercase , transformer=_lowercase , text_encoder=_lowercase , tokenizer=_lowercase , scheduler=_lowercase , learned_classifier_free_sampling_embeddings=_lowercase , ) def _lowercase ( self , _lowercase , _lowercase , _lowercase ): """simple docstring""" _lowerCAmelCase = len(_lowercase ) if isinstance(_lowercase , _lowercase ) else 1 # get prompt text embeddings _lowerCAmelCase = self.tokenizer( _lowercase , padding="""max_length""" , max_length=self.tokenizer.model_max_length , return_tensors="""pt""" , ) _lowerCAmelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _lowerCAmelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F' {self.tokenizer.model_max_length} tokens: {removed_text}' ) _lowerCAmelCase = text_input_ids[:, : self.tokenizer.model_max_length] _lowerCAmelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 _lowerCAmelCase = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=_lowercase ) # duplicate text embeddings for each generation per prompt _lowerCAmelCase = prompt_embeds.repeat_interleave(_lowercase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: _lowerCAmelCase = self.learned_classifier_free_sampling_embeddings.embeddings _lowerCAmelCase = negative_prompt_embeds.unsqueeze(0 ).repeat(_lowercase , 1 , 1 ) else: _lowerCAmelCase = [""""""] * batch_size _lowerCAmelCase = text_input_ids.shape[-1] _lowerCAmelCase = self.tokenizer( _lowercase , padding="""max_length""" , max_length=_lowercase , truncation=_lowercase , return_tensors="""pt""" , ) _lowerCAmelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings _lowerCAmelCase = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=_lowercase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _lowerCAmelCase = negative_prompt_embeds.shape[1] _lowerCAmelCase = negative_prompt_embeds.repeat(1 , _lowercase , 1 ) _lowerCAmelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , _lowercase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _lowerCAmelCase = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , _lowercase , _lowercase = 100 , _lowercase = 5.0 , _lowercase = 1.0 , _lowercase = 1 , _lowercase = None , _lowercase = None , _lowercase = "pil" , _lowercase = True , _lowercase = None , _lowercase = 1 , ): """simple docstring""" if isinstance(_lowercase , _lowercase ): _lowerCAmelCase = 1 elif isinstance(_lowercase , _lowercase ): _lowerCAmelCase = len(_lowercase ) else: raise ValueError(F'`prompt` has to be of type `str` or `list` but is {type(_lowercase )}' ) _lowerCAmelCase = batch_size * num_images_per_prompt _lowerCAmelCase = guidance_scale > 1.0 _lowerCAmelCase = self._encode_prompt(_lowercase , _lowercase , _lowercase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_lowercase , _lowercase ) or callback_steps <= 0) ): raise ValueError( F'`callback_steps` has to be a positive integer but is {callback_steps} of type' F' {type(_lowercase )}.' ) # get the initial completely masked latents unless the user supplied it _lowerCAmelCase = (batch_size, self.transformer.num_latent_pixels) if latents is None: _lowerCAmelCase = self.transformer.num_vector_embeds - 1 _lowerCAmelCase = torch.full(_lowercase , _lowercase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {latents_shape}' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( """Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,""" F' {self.transformer.num_vector_embeds - 1} (inclusive).' ) _lowerCAmelCase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(_lowercase , device=self.device ) _lowerCAmelCase = self.scheduler.timesteps.to(self.device ) _lowerCAmelCase = latents for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the sample if we are doing classifier free guidance _lowerCAmelCase = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` _lowerCAmelCase = self.transformer(_lowercase , encoder_hidden_states=_lowercase , timestep=_lowercase ).sample if do_classifier_free_guidance: _lowerCAmelCase , _lowerCAmelCase = model_output.chunk(2 ) _lowerCAmelCase = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(_lowercase , dim=1 , keepdim=_lowercase ) _lowerCAmelCase = self.truncate(_lowercase , _lowercase ) # remove `log(0)`'s (`-inf`s) _lowerCAmelCase = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 _lowerCAmelCase = self.scheduler.step(_lowercase , timestep=_lowercase , sample=_lowercase , generator=_lowercase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_lowercase , _lowercase , _lowercase ) _lowerCAmelCase = self.vqvae.config.vq_embed_dim _lowerCAmelCase = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) _lowerCAmelCase = self.vqvae.quantize.get_codebook_entry(_lowercase , shape=_lowercase ) _lowerCAmelCase = self.vqvae.decode(_lowercase , force_not_quantize=_lowercase ).sample _lowerCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _lowerCAmelCase = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase ) def _lowercase ( self , _lowercase , _lowercase ): """simple docstring""" _lowerCAmelCase , _lowerCAmelCase = torch.sort(_lowercase , 1 , descending=_lowercase ) _lowerCAmelCase = torch.exp(_lowercase ) _lowerCAmelCase = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out _lowerCAmelCase = torch.full_like(keep_mask[:, 0:1, :] , _lowercase ) _lowerCAmelCase = torch.cat((all_true, keep_mask) , dim=1 ) _lowerCAmelCase = keep_mask[:, :-1, :] _lowerCAmelCase = keep_mask.gather(1 , indices.argsort(1 ) ) _lowerCAmelCase = log_p_x_0.clone() _lowerCAmelCase = -torch.inf # -inf = log(0) return rv
229
1
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def lowercase__ ( __snake_case : Tuple , __snake_case : Tuple=None ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = None if token is not None: UpperCAmelCase_ : Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F"Bearer {token}"} UpperCAmelCase_ : List[str] = F"https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100" UpperCAmelCase_ : Any = requests.get(__snake_case , headers=__snake_case ).json() UpperCAmelCase_ : List[str] = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) UpperCAmelCase_ : Any = math.ceil((result['total_count'] - 100) / 100 ) for i in range(__snake_case ): UpperCAmelCase_ : Union[str, Any] = requests.get(url + F"&page={i + 2}" , headers=__snake_case ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F"Unknown error, could not fetch links:\n{traceback.format_exc()}" ) return {} def lowercase__ ( __snake_case : Optional[int] , __snake_case : Tuple=None ): '''simple docstring''' UpperCAmelCase_ : Dict = None if token is not None: UpperCAmelCase_ : List[Any] = {'Accept': 'application/vnd.github+json', 'Authorization': F"Bearer {token}"} UpperCAmelCase_ : Dict = F"https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100" UpperCAmelCase_ : int = requests.get(__snake_case , headers=__snake_case ).json() UpperCAmelCase_ : Optional[int] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) UpperCAmelCase_ : Optional[Any] = math.ceil((result['total_count'] - 100) / 100 ) for i in range(__snake_case ): UpperCAmelCase_ : Dict = requests.get(url + F"&page={i + 2}" , headers=__snake_case ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F"Unknown error, could not fetch links:\n{traceback.format_exc()}" ) return {} def lowercase__ ( __snake_case : str , __snake_case : str , __snake_case : int , __snake_case : Dict ): '''simple docstring''' UpperCAmelCase_ : Any = None if token is not None: UpperCAmelCase_ : Any = {'Accept': 'application/vnd.github+json', 'Authorization': F"Bearer {token}"} UpperCAmelCase_ : Any = requests.get(__snake_case , headers=__snake_case , allow_redirects=__snake_case ) UpperCAmelCase_ : List[Any] = result.headers['Location'] UpperCAmelCase_ : str = requests.get(__snake_case , allow_redirects=__snake_case ) UpperCAmelCase_ : str = os.path.join(__snake_case , F"{artifact_name}.zip" ) with open(__snake_case , 'wb' ) as fp: fp.write(response.content ) def lowercase__ ( __snake_case : Any , __snake_case : Any=None ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : List[Any] = [] UpperCAmelCase_ : List[str] = None with zipfile.ZipFile(__snake_case ) as z: for filename in z.namelist(): if not os.path.isdir(__snake_case ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(__snake_case ) as f: for line in f: UpperCAmelCase_ : List[str] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs UpperCAmelCase_ : Optional[Any] = line[: line.index(': ' )] UpperCAmelCase_ : Dict = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed UpperCAmelCase_ : List[Any] = line[len('FAILED ' ) :] failed_tests.append(__snake_case ) elif filename == "job_name.txt": UpperCAmelCase_ : Optional[Any] = line if len(__snake_case ) != len(__snake_case ): raise ValueError( F"`errors` and `failed_tests` should have the same number of elements. Got {len(__snake_case )} for `errors` " F"and {len(__snake_case )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some" ' problem.' ) UpperCAmelCase_ : List[str] = None if job_name and job_links: UpperCAmelCase_ : Optional[int] = job_links.get(__snake_case , __snake_case ) # A list with elements of the form (line of error, error, failed test) UpperCAmelCase_ : Union[str, Any] = [x + [y] + [job_link] for x, y in zip(__snake_case , __snake_case )] return result def lowercase__ ( __snake_case : str , __snake_case : Union[str, Any]=None ): '''simple docstring''' UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : int = [os.path.join(__snake_case , __snake_case ) for p in os.listdir(__snake_case ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(__snake_case , job_links=__snake_case ) ) return errors def lowercase__ ( __snake_case : int , __snake_case : str=None ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = Counter() counter.update([x[1] for x in logs] ) UpperCAmelCase_ : List[str] = counter.most_common() UpperCAmelCase_ : int = {} for error, count in counts: if error_filter is None or error not in error_filter: UpperCAmelCase_ : List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} UpperCAmelCase_ : Union[str, Any] = dict(sorted(r.items() , key=lambda __snake_case : item[1]["count"] , reverse=__snake_case ) ) return r def lowercase__ ( __snake_case : Optional[Any] ): '''simple docstring''' UpperCAmelCase_ : Tuple = test.split('::' )[0] if test.startswith('tests/models/' ): UpperCAmelCase_ : Optional[int] = test.split('/' )[2] else: UpperCAmelCase_ : Tuple = None return test def lowercase__ ( __snake_case : Optional[Any] , __snake_case : int=None ): '''simple docstring''' UpperCAmelCase_ : Any = [(x[0], x[1], get_model(x[2] )) for x in logs] UpperCAmelCase_ : Dict = [x for x in logs if x[2] is not None] UpperCAmelCase_ : List[str] = {x[2] for x in logs} UpperCAmelCase_ : int = {} for test in tests: UpperCAmelCase_ : int = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) UpperCAmelCase_ : Optional[Any] = counter.most_common() UpperCAmelCase_ : Union[str, Any] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} UpperCAmelCase_ : Union[str, Any] = sum(error_counts.values() ) if n_errors > 0: UpperCAmelCase_ : List[Any] = {'count': n_errors, 'errors': error_counts} UpperCAmelCase_ : List[str] = dict(sorted(r.items() , key=lambda __snake_case : item[1]["count"] , reverse=__snake_case ) ) return r def lowercase__ ( __snake_case : Dict ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = '| no. | error | status |' UpperCAmelCase_ : str = '|-:|:-|:-|' UpperCAmelCase_ : Optional[int] = [header, sep] for error in reduced_by_error: UpperCAmelCase_ : Union[str, Any] = reduced_by_error[error]['count'] UpperCAmelCase_ : Optional[Any] = F"| {count} | {error[:100]} | |" lines.append(__snake_case ) return "\n".join(__snake_case ) def lowercase__ ( __snake_case : str ): '''simple docstring''' UpperCAmelCase_ : Dict = '| model | no. of errors | major error | count |' UpperCAmelCase_ : Union[str, Any] = '|-:|-:|-:|-:|' UpperCAmelCase_ : Any = [header, sep] for model in reduced_by_model: UpperCAmelCase_ : Tuple = reduced_by_model[model]['count'] UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = list(reduced_by_model[model]['errors'].items() )[0] UpperCAmelCase_ : Any = F"| {model} | {count} | {error[:60]} | {_count} |" lines.append(__snake_case ) return "\n".join(__snake_case ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') __UpperCAmelCase = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) __UpperCAmelCase = get_job_links(args.workflow_run_id, token=args.token) __UpperCAmelCase = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: __UpperCAmelCase = k.find(' / ') __UpperCAmelCase = k[index + len(' / ') :] __UpperCAmelCase = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) __UpperCAmelCase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) __UpperCAmelCase = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error __UpperCAmelCase = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors __UpperCAmelCase = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) __UpperCAmelCase = reduce_by_error(errors) __UpperCAmelCase = reduce_by_model(errors) __UpperCAmelCase = make_github_table(reduced_by_error) __UpperCAmelCase = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
29
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase : Any = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[str] = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys UpperCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
95
0
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants _UpperCamelCase = Mapping[str, np.ndarray] _UpperCamelCase = Mapping[str, Any] # Is a nested dict. _UpperCamelCase = 0.0_1 @dataclasses.dataclass(frozen=_UpperCamelCase ) class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. __SCREAMING_SNAKE_CASE = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. __SCREAMING_SNAKE_CASE = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. __SCREAMING_SNAKE_CASE = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. __SCREAMING_SNAKE_CASE = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions __SCREAMING_SNAKE_CASE = None # Optional remark about the protein. Included as a comment in output PDB # files __SCREAMING_SNAKE_CASE = None # Templates used to generate this protein (prediction-only) __SCREAMING_SNAKE_CASE = None # Chain corresponding to each parent __SCREAMING_SNAKE_CASE = None def UpperCamelCase_( snake_case__: str ) -> Protein: UpperCAmelCase__ = r'(\[[A-Z]+\]\n)' UpperCAmelCase__ = [tag.strip() for tag in re.split(snake_case__ , snake_case__ ) if len(snake_case__ ) > 0] UpperCAmelCase__ = zip(tags[0::2] , [l.split('\n' ) for l in tags[1::2]] ) UpperCAmelCase__ = ["N", "CA", "C"] UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None for g in groups: if "[PRIMARY]" == g[0]: UpperCAmelCase__ = g[1][0].strip() for i in range(len(snake_case__ ) ): if seq[i] not in residue_constants.restypes: UpperCAmelCase__ = 'X' # FIXME: strings are immutable UpperCAmelCase__ = np.array( [residue_constants.restype_order.get(snake_case__ , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: UpperCAmelCase__ = [] for axis in range(3 ): tertiary.append(list(map(snake_case__ , g[1][axis].split() ) ) ) UpperCAmelCase__ = np.array(snake_case__ ) UpperCAmelCase__ = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(snake_case__ ): UpperCAmelCase__ = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: UpperCAmelCase__ = np.array(list(map({'-': 0, '+': 1}.get , g[1][0].strip() ) ) ) UpperCAmelCase__ = np.zeros( ( len(snake_case__ ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(snake_case__ ): UpperCAmelCase__ = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=snake_case__ , atom_mask=snake_case__ , aatype=snake_case__ , residue_index=np.arange(len(snake_case__ ) ) , b_factors=snake_case__ , ) def UpperCamelCase_( snake_case__: Protein , snake_case__: int = 0 ) -> List[str]: UpperCAmelCase__ = [] UpperCAmelCase__ = prot.remark if remark is not None: pdb_headers.append(f"REMARK {remark}" ) UpperCAmelCase__ = prot.parents UpperCAmelCase__ = prot.parents_chain_index if parents is not None and parents_chain_index is not None: UpperCAmelCase__ = [p for i, p in zip(snake_case__ , snake_case__ ) if i == chain_id] if parents is None or len(snake_case__ ) == 0: UpperCAmelCase__ = ['N/A'] pdb_headers.append(f"PARENT {' '.join(snake_case__ )}" ) return pdb_headers def UpperCamelCase_( snake_case__: Protein , snake_case__: str ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = pdb_str.split('\n' ) UpperCAmelCase__ = prot.remark if remark is not None: out_pdb_lines.append(f"REMARK {remark}" ) UpperCAmelCase__ = 42 if prot.parents is not None and len(prot.parents ) > 0: UpperCAmelCase__ = [] if prot.parents_chain_index is not None: UpperCAmelCase__ = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(snake_case__ ) , [] ) parent_dict[str(snake_case__ )].append(snake_case__ ) UpperCAmelCase__ = max([int(snake_case__ ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): UpperCAmelCase__ = parent_dict.get(str(snake_case__ ) , ['N/A'] ) parents_per_chain.append(snake_case__ ) else: parents_per_chain.append(list(prot.parents ) ) else: UpperCAmelCase__ = [['N/A']] def make_parent_line(snake_case__: Sequence[str] ) -> str: return f"PARENT {' '.join(snake_case__ )}" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) UpperCAmelCase__ = 0 for i, l in enumerate(snake_case__ ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(snake_case__ ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(snake_case__ ): UpperCAmelCase__ = parents_per_chain[chain_counter] else: UpperCAmelCase__ = ['N/A'] out_pdb_lines.append(make_parent_line(snake_case__ ) ) return "\n".join(snake_case__ ) def UpperCamelCase_( snake_case__: Protein ) -> str: UpperCAmelCase__ = residue_constants.restypes + ['X'] def res_atoa(snake_case__: int ) -> str: return residue_constants.restype_atoa.get(restypes[r] , 'UNK' ) UpperCAmelCase__ = residue_constants.atom_types UpperCAmelCase__ = [] UpperCAmelCase__ = prot.atom_mask UpperCAmelCase__ = prot.aatype UpperCAmelCase__ = prot.atom_positions UpperCAmelCase__ = prot.residue_index.astype(np.intaa ) UpperCAmelCase__ = prot.b_factors UpperCAmelCase__ = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError('Invalid aatypes.' ) UpperCAmelCase__ = get_pdb_headers(snake_case__ ) if len(snake_case__ ) > 0: pdb_lines.extend(snake_case__ ) UpperCAmelCase__ = aatype.shape[0] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 UpperCAmelCase__ = string.ascii_uppercase UpperCAmelCase__ = None # Add all atom sites. for i in range(snake_case__ ): UpperCAmelCase__ = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(snake_case__ , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue UpperCAmelCase__ = 'ATOM' UpperCAmelCase__ = atom_name if len(snake_case__ ) == 4 else f" {atom_name}" UpperCAmelCase__ = '' UpperCAmelCase__ = '' UpperCAmelCase__ = 1.0_0 UpperCAmelCase__ = atom_name[0] # Protein supports only C, N, O, S, this works. UpperCAmelCase__ = '' UpperCAmelCase__ = 'A' if chain_index is not None: UpperCAmelCase__ = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! UpperCAmelCase__ = ( f"{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}" f"{res_name_a:>3} {chain_tag:>1}" f"{residue_index[i]:>4}{insertion_code:>1} " f"{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}" f"{occupancy:>6.2f}{b_factor:>6.2f} " f"{element:>2}{charge:>2}" ) pdb_lines.append(snake_case__ ) atom_index += 1 UpperCAmelCase__ = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: UpperCAmelCase__ = True UpperCAmelCase__ = chain_index[i + 1] if should_terminate: # Close the chain. UpperCAmelCase__ = 'TER' UpperCAmelCase__ = ( f"{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}" ) pdb_lines.append(snake_case__ ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(snake_case__ , snake_case__ ) ) pdb_lines.append('END' ) pdb_lines.append('' ) return "\n".join(snake_case__ ) def UpperCamelCase_( snake_case__: Protein ) -> np.ndarray: return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def UpperCamelCase_( snake_case__: FeatureDict , snake_case__: ModelOutput , snake_case__: Optional[np.ndarray] = None , snake_case__: Optional[np.ndarray] = None , snake_case__: Optional[str] = None , snake_case__: Optional[Sequence[str]] = None , snake_case__: Optional[Sequence[int]] = None , ) -> Protein: return Protein( aatype=features['aatype'] , atom_positions=result['final_atom_positions'] , atom_mask=result['final_atom_mask'] , residue_index=features['residue_index'] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result['final_atom_mask'] ) , chain_index=snake_case__ , remark=snake_case__ , parents=snake_case__ , parents_chain_index=snake_case__ , )
335
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class lowercase ( unittest.TestCase ): '''simple docstring''' def __init__(self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=16 , __a=2 , __a=0.02 , __a=4 , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_choices def UpperCamelCase__ (self ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_attention_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = True UpperCAmelCase__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowercase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ = FlaxRobertaModelTester(self ) @slow def UpperCamelCase__ (self ) -> str: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase__ = model_class_name.from_pretrained('roberta-base' , from_pt=__a ) UpperCAmelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__a )
335
1
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCAmelCase__ = random.Random() def _UpperCAmelCase (UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[int]=1.0 , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Optional[Any]=None ): if rng is None: _A : Dict = global_rng _A : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCAmelCase__ ( unittest.TestCase): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase=7 , __lowerCamelCase=4_0_0 , __lowerCamelCase=2_0_0_0 , __lowerCamelCase=2_4 , __lowerCamelCase=2_4 , __lowerCamelCase=0.0 , __lowerCamelCase=1_6_0_0_0 , __lowerCamelCase=True , __lowerCamelCase=True , ) -> Tuple: _A : Tuple = parent _A : Any = batch_size _A : List[Any] = min_seq_length _A : List[Any] = max_seq_length _A : int = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _A : Optional[Any] = feature_size _A : List[Any] = num_mel_bins _A : Optional[int] = padding_value _A : List[Any] = sampling_rate _A : List[Any] = return_attention_mask _A : List[str] = do_normalize def _lowerCamelCase ( self) -> List[Any]: return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowerCamelCase ( self , __lowerCamelCase=False , __lowerCamelCase=False) -> Union[str, Any]: def _flatten(__lowerCamelCase): return list(itertools.chain(*__lowerCamelCase)) if equal_length: _A : List[Any] = [floats_list((self.max_seq_length, self.feature_size)) for _ in range(self.batch_size)] else: # make sure that inputs increase in size _A : List[Any] = [ floats_list((x, self.feature_size)) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: _A : str = [np.asarray(__lowerCamelCase) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCAmelCase__ ( a , unittest.TestCase): '''simple docstring''' __SCREAMING_SNAKE_CASE = SpeechaTextFeatureExtractor if is_speech_available() else None def _lowerCamelCase ( self) -> Any: _A : Dict = SpeechaTextFeatureExtractionTester(self) def _lowerCamelCase ( self , __lowerCamelCase) -> Any: self.assertTrue(np.all(np.mean(__lowerCamelCase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(__lowerCamelCase , axis=0) - 1) < 1e-3)) def _lowerCamelCase ( self) -> Dict: # Tests that all call wrap to encode_plus and batch_encode_plus _A : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 _A : List[str] = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : Any = [np.asarray(__lowerCamelCase) for speech_input in speech_inputs] # Test feature size _A : List[Any] = feature_extractor(__lowerCamelCase , padding=__lowerCamelCase , return_tensors="np").input_features self.assertTrue(input_features.ndim == 3) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size) # Test not batched input _A : Optional[int] = feature_extractor(speech_inputs[0] , return_tensors="np").input_features _A : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np").input_features self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3)) # Test batched _A : Optional[int] = feature_extractor(__lowerCamelCase , return_tensors="np").input_features _A : Optional[int] = feature_extractor(__lowerCamelCase , return_tensors="np").input_features for enc_seq_a, enc_seq_a in zip(__lowerCamelCase , __lowerCamelCase): self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3)) # Test 2-D numpy arrays are batched. _A : int = [floats_list((1, x))[0] for x in (8_0_0, 8_0_0, 8_0_0)] _A : Optional[Any] = np.asarray(__lowerCamelCase) _A : Dict = feature_extractor(__lowerCamelCase , return_tensors="np").input_features _A : Union[str, Any] = feature_extractor(__lowerCamelCase , return_tensors="np").input_features for enc_seq_a, enc_seq_a in zip(__lowerCamelCase , __lowerCamelCase): self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3)) def _lowerCamelCase ( self) -> Dict: _A : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : int = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : int = ["longest", "max_length", "do_not_pad"] _A : int = [None, 1_6, None] for max_length, padding in zip(__lowerCamelCase , __lowerCamelCase): _A : Optional[Any] = feature_extractor( __lowerCamelCase , padding=__lowerCamelCase , max_length=__lowerCamelCase , return_attention_mask=__lowerCamelCase) _A : Union[str, Any] = inputs.input_features _A : int = inputs.attention_mask _A : List[str] = [np.sum(__lowerCamelCase) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]]) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]]) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]]) def _lowerCamelCase ( self) -> Optional[int]: _A : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : int = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : Any = ["longest", "max_length", "do_not_pad"] _A : str = [None, 1_6, None] for max_length, padding in zip(__lowerCamelCase , __lowerCamelCase): _A : Any = feature_extractor( __lowerCamelCase , max_length=__lowerCamelCase , padding=__lowerCamelCase , return_tensors="np" , return_attention_mask=__lowerCamelCase) _A : Dict = inputs.input_features _A : str = inputs.attention_mask _A : int = [np.sum(__lowerCamelCase) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]]) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1e-6) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]]) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1e-6) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]]) def _lowerCamelCase ( self) -> Dict: _A : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : Optional[int] = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : Tuple = feature_extractor( __lowerCamelCase , padding="max_length" , max_length=4 , truncation=__lowerCamelCase , return_tensors="np" , return_attention_mask=__lowerCamelCase , ) _A : Tuple = inputs.input_features _A : Optional[int] = inputs.attention_mask _A : Optional[Any] = np.sum(attention_mask == 1 , axis=1) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]]) self._check_zero_mean_unit_variance(input_features[1]) self._check_zero_mean_unit_variance(input_features[2]) def _lowerCamelCase ( self) -> Dict: _A : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : Union[str, Any] = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : Optional[int] = feature_extractor( __lowerCamelCase , padding="longest" , max_length=4 , truncation=__lowerCamelCase , return_tensors="np" , return_attention_mask=__lowerCamelCase , ) _A : List[Any] = inputs.input_features _A : int = inputs.attention_mask _A : Tuple = np.sum(attention_mask == 1 , axis=1) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]]) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]]) self._check_zero_mean_unit_variance(input_features[2]) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 2_4)) _A : List[str] = [floats_list((1, x))[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0)] _A : List[Any] = feature_extractor( __lowerCamelCase , padding="longest" , max_length=1_6 , truncation=__lowerCamelCase , return_tensors="np" , return_attention_mask=__lowerCamelCase , ) _A : Optional[int] = inputs.input_features _A : Tuple = inputs.attention_mask _A : List[str] = np.sum(attention_mask == 1 , axis=1) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]]) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]]) self._check_zero_mean_unit_variance(input_features[2]) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 2_4)) def _lowerCamelCase ( self) -> str: import torch _A : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : str = np.random.rand(1_0_0 , 3_2).astype(np.floataa) _A : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _A : Dict = feature_extractor.pad([{"input_features": inputs}] , return_tensors="np") self.assertTrue(np_processed.input_features.dtype == np.floataa) _A : Dict = feature_extractor.pad([{"input_features": inputs}] , return_tensors="pt") self.assertTrue(pt_processed.input_features.dtype == torch.floataa) def _lowerCamelCase ( self , __lowerCamelCase) -> str: from datasets import load_dataset _A : Union[str, Any] = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation") # automatic decoding with librispeech _A : Dict = ds.sort("id").select(range(__lowerCamelCase))[:num_samples]["audio"] return [x["array"] for x in speech_samples] def _lowerCamelCase ( self) -> Any: # fmt: off _A : Dict = np.array([ -1.5_7_4_5, -1.7_7_1_3, -1.7_0_2_0, -1.6_0_6_9, -1.2_2_5_0, -1.1_1_0_5, -0.9_0_7_2, -0.8_2_4_1, -1.2_3_1_0, -0.8_0_9_8, -0.3_3_2_0, -0.4_1_0_1, -0.7_9_8_5, -0.4_9_9_6, -0.8_2_1_3, -0.9_1_2_8, -1.0_4_2_0, -1.1_2_8_6, -1.0_4_4_0, -0.7_9_9_9, -0.8_4_0_5, -1.2_2_7_5, -1.5_4_4_3, -1.4_6_2_5, ]) # fmt: on _A : Union[str, Any] = self._load_datasamples(1) _A : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) _A : Tuple = feature_extractor(__lowerCamelCase , return_tensors="pt").input_features self.assertEquals(input_features.shape , (1, 5_8_4, 2_4)) self.assertTrue(np.allclose(input_features[0, 0, :3_0] , __lowerCamelCase , atol=1e-4))
11
'''simple docstring''' from __future__ import annotations import queue class lowercase : """simple docstring""" def __init__( self ,a_ ) -> str: _UpperCAmelCase : Optional[Any] = data _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Union[str, Any] = None def snake_case_ ( )-> TreeNode: '''simple docstring''' print("""\n********Press N to stop entering at any point of time********\n""" ) _UpperCAmelCase : Any = input("""Enter the value of the root node: """ ).strip().lower() _UpperCAmelCase : queue.Queue = queue.Queue() _UpperCAmelCase : List[str] = TreeNode(int(lowerCAmelCase_ ) ) q.put(lowerCAmelCase_ ) while not q.empty(): _UpperCAmelCase : str = q.get() _UpperCAmelCase : Any = F'''Enter the left node of {node_found.data}: ''' _UpperCAmelCase : Union[str, Any] = input(lowerCAmelCase_ ).strip().lower() or """n""" if check == "n": return tree_node _UpperCAmelCase : List[str] = TreeNode(int(lowerCAmelCase_ ) ) _UpperCAmelCase : Optional[int] = left_node q.put(lowerCAmelCase_ ) _UpperCAmelCase : Dict = F'''Enter the right node of {node_found.data}: ''' _UpperCAmelCase : Tuple = input(lowerCAmelCase_ ).strip().lower() or """n""" if check == "n": return tree_node _UpperCAmelCase : Any = TreeNode(int(lowerCAmelCase_ ) ) _UpperCAmelCase : Optional[Any] = right_node q.put(lowerCAmelCase_ ) raise def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return print(node.data , end=""",""" ) pre_order(node.left ) pre_order(node.right ) def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return in_order(node.left ) print(node.data , end=""",""" ) in_order(node.right ) def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=""",""" ) def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return _UpperCAmelCase : queue.Queue = queue.Queue() q.put(lowerCAmelCase_ ) while not q.empty(): _UpperCAmelCase : Dict = q.get() print(node_dequeued.data , end=""",""" ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return _UpperCAmelCase : queue.Queue = queue.Queue() q.put(lowerCAmelCase_ ) while not q.empty(): _UpperCAmelCase : Optional[int] = [] while not q.empty(): _UpperCAmelCase : Optional[int] = q.get() print(node_dequeued.data , end=""",""" ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(lowerCAmelCase_ ) def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return _UpperCAmelCase : list[TreeNode] = [] _UpperCAmelCase : Optional[Any] = node while n or stack: while n: # start from root node, find its left child print(n.data , end=""",""" ) stack.append(lowerCAmelCase_ ) _UpperCAmelCase : Union[str, Any] = n.left # end of while means current node doesn't have left child _UpperCAmelCase : int = stack.pop() # start to traverse its right child _UpperCAmelCase : Any = n.right def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return _UpperCAmelCase : list[TreeNode] = [] _UpperCAmelCase : Optional[Any] = node while n or stack: while n: stack.append(lowerCAmelCase_ ) _UpperCAmelCase : Tuple = n.left _UpperCAmelCase : Union[str, Any] = stack.pop() print(n.data , end=""",""" ) _UpperCAmelCase : Any = n.right def snake_case_ ( lowerCAmelCase_ )-> None: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not node: return _UpperCAmelCase ,_UpperCAmelCase : str = [], [] _UpperCAmelCase : Dict = node stacka.append(lowerCAmelCase_ ) while stacka: # to find the reversed order of post order, store it in stack2 _UpperCAmelCase : Optional[int] = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(lowerCAmelCase_ ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=""",""" ) def snake_case_ ( lowerCAmelCase_ = "" , lowerCAmelCase_=50 , lowerCAmelCase_="*" )-> str: '''simple docstring''' if not s: return "\n" + width * char _UpperCAmelCase ,_UpperCAmelCase : Optional[Any] = divmod(width - len(lowerCAmelCase_ ) - 2 , 2 ) return F'''{left * char} {s} {(left + extra) * char}''' if __name__ == "__main__": import doctest doctest.testmod() print(prompt("""Binary Tree Traversals""")) A_ : TreeNode = build_tree() print(prompt("""Pre Order Traversal""")) pre_order(node) print(prompt() + """\n""") print(prompt("""In Order Traversal""")) in_order(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal""")) post_order(node) print(prompt() + """\n""") print(prompt("""Level Order Traversal""")) level_order(node) print(prompt() + """\n""") print(prompt("""Actual Level Order Traversal""")) level_order_actual(node) print("""*""" * 5_0 + """\n""") print(prompt("""Pre Order Traversal - Iteration Version""")) pre_order_iter(node) print(prompt() + """\n""") print(prompt("""In Order Traversal - Iteration Version""")) in_order_iter(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal - Iteration Version""")) post_order_iter(node) print(prompt())
215
0
from collections import deque from .hash_table import HashTable class lowercase ( UpperCamelCase__ ): def __init__( self , *_a , **_a ) -> Optional[int]: super().__init__(*_a , **_a ) def a__ ( self , _a , _a ) -> Union[str, Any]: _A : int = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_a ) _A : Tuple = self.values[key] def a__ ( self ) -> int: return ( sum(self.charge_factor - len(_a ) for slot in self.values ) / self.size_table * self.charge_factor ) def a__ ( self , _a , _a=None ) -> Dict: if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_a ) == 0 ): return key return super()._collision_resolution(_a , _a )
343
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowercase : _a = 42 # setable values _a = 42 _a = 42 _a = None @classmethod def a__ ( cls , _a , _a , _a ) -> Tuple: return cls(common=_a , init_noise_sigma=_a , timesteps=_a ) @dataclass class lowercase ( UpperCamelCase__ ): _a = 42 class lowercase ( UpperCamelCase__,UpperCamelCase__ ): _a = [e.name for e in FlaxKarrasDiffusionSchedulers] _a = 42 @property def a__ ( self ) -> Dict: return True @register_to_config def __init__( self , _a = 1000 , _a = 0.0001 , _a = 0.02 , _a = "linear" , _a = None , _a = "fixed_small" , _a = True , _a = "epsilon" , _a = jnp.floataa , ) -> Tuple: _A : Tuple = dtype def a__ ( self , _a = None ) -> DDPMSchedulerState: if common is None: _A : Dict = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution _A : Union[str, Any] = jnp.array(1.0 , dtype=self.dtype ) _A : Tuple = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_a , init_noise_sigma=_a , timesteps=_a , ) def a__ ( self , _a , _a , _a = None ) -> jnp.ndarray: return sample def a__ ( self , _a , _a , _a = () ) -> DDPMSchedulerState: _A : Any = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 _A : Dict = (jnp.arange(0 , _a ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_a , timesteps=_a , ) def a__ ( self , _a , _a , _a=None , _a=None ) -> Optional[int]: _A : Optional[Any] = state.common.alphas_cumprod[t] _A : int = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _A : List[str] = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: _A : Optional[Any] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": _A : Optional[Any] = jnp.clip(_a , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": _A : Any = jnp.log(jnp.clip(_a , a_min=1e-20 ) ) elif variance_type == "fixed_large": _A : Optional[Any] = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log _A : Tuple = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": _A : str = variance _A : Union[str, Any] = state.common.betas[t] _A : Tuple = (predicted_variance + 1) / 2 _A : List[str] = frac * max_log + (1 - frac) * min_log return variance def a__ ( self , _a , _a , _a , _a , _a = None , _a = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: _A : Dict = timestep if key is None: _A : int = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: _A , _A : List[str] = jnp.split(_a , sample.shape[1] , axis=1 ) else: _A : int = None # 1. compute alphas, betas _A : int = state.common.alphas_cumprod[t] _A : List[str] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) _A : Union[str, Any] = 1 - alpha_prod_t _A : Optional[int] = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _A : Dict = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _A : Optional[int] = model_output elif self.config.prediction_type == "v_prediction": _A : Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' """ for the FlaxDDPMScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: _A : Union[str, Any] = jnp.clip(_a , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _A : List[Any] = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t _A : Dict = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _A : int = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): _A : Tuple = jax.random.split(_a , num=1 ) _A : Dict = jax.random.normal(_a , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_a , _a , predicted_variance=_a ) ** 0.5) * noise _A : int = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) _A : Union[str, Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_a , state=_a ) def a__ ( self , _a , _a , _a , _a , ) -> jnp.ndarray: return add_noise_common(state.common , _a , _a , _a ) def a__ ( self , _a , _a , _a , _a , ) -> jnp.ndarray: return get_velocity_common(state.common , _a , _a , _a ) def __len__( self ) -> List[Any]: return self.config.num_train_timesteps
343
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _A = logging.get_logger(__name__) _A = '▁' _A = {'vocab_file': 'sentencepiece.bpe.model', 'monolingual_vocab_file': 'dict.txt'} _A = { 'vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model', }, 'monolingual_vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt', }, } _A = {'vinai/bartpho-syllable': 1024} class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Any = VOCAB_FILES_NAMES UpperCAmelCase__ : int = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : str = ["input_ids", "attention_mask"] def __init__( self , A_ , A_ , A_="<s>" , A_="</s>" , A_="</s>" , A_="<s>" , A_="<unk>" , A_="<pad>" , A_="<mask>" , A_ = None , **A_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it __UpperCamelCase =AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else mask_token __UpperCamelCase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A_ , eos_token=A_ , unk_token=A_ , sep_token=A_ , cls_token=A_ , pad_token=A_ , mask_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __UpperCamelCase =vocab_file __UpperCamelCase =monolingual_vocab_file __UpperCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(A_ ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility __UpperCamelCase ={} __UpperCamelCase =0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(A_ ) not in self.fairseq_tokens_to_ids: __UpperCamelCase =cnt cnt += 1 with open(A_ , 'r' , encoding='utf-8' ) as f: for line in f.readlines(): __UpperCamelCase =line.strip().split()[0] __UpperCamelCase =len(self.fairseq_tokens_to_ids ) if str(A_ ) not in self.fairseq_tokens_to_ids: __UpperCamelCase =len(self.fairseq_tokens_to_ids ) __UpperCamelCase ={v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> Any: __UpperCamelCase =self.__dict__.copy() __UpperCamelCase =None __UpperCamelCase =self.sp_model.serialized_model_proto() return state def __setstate__( self , A_ ) -> List[str]: __UpperCamelCase =d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __UpperCamelCase ={} __UpperCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _a ( self , A_ , A_ = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __UpperCamelCase =[self.cls_token_id] __UpperCamelCase =[self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _a ( self , A_ , A_ = None , A_ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is None: return [1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1, 1] + ([0] * len(A_ )) + [1] def _a ( self , A_ , A_ = None ) -> List[int]: __UpperCamelCase =[self.sep_token_id] __UpperCamelCase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _a ( self ) -> Any: return len(self.fairseq_ids_to_tokens ) def _a ( self ) -> Union[str, Any]: __UpperCamelCase ={self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _a ( self , A_ ) -> List[str]: return self.sp_model.encode(A_ , out_type=A_ ) def _a ( self , A_ ) -> str: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def _a ( self , A_ ) -> int: return self.fairseq_ids_to_tokens[index] def _a ( self , A_ ) -> List[Any]: __UpperCamelCase =''.join(A_ ).replace(A_ , ' ' ).strip() return out_string def _a ( self , A_ , A_ = None ) -> Tuple[str]: if not os.path.isdir(A_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __UpperCamelCase =os.path.join( A_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase =os.path.join( A_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , 'wb' ) as fi: __UpperCamelCase =self.sp_model.serialized_model_proto() fi.write(A_ ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( A_ ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , A_ ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(A_ , 'w' , encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f'{str(A_ )} \n' ) return out_vocab_file, out_monolingual_vocab_file
62
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( *lowerCamelCase__ : List[Any] ,**lowerCamelCase__ : Union[str, Any] ): pass def a_ ( lowerCamelCase ): UpperCAmelCase__ = hashlib.mda(image.tobytes() ) return m.hexdigest()[:1_0] def a_ ( lowerCamelCase ): UpperCAmelCase__ = np.array(lowerCamelCase ) UpperCAmelCase__ = npimg.shape return {"hash": hashimage(lowerCamelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class snake_case ( unittest.TestCase ): """simple docstring""" snake_case__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) snake_case__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : str ): UpperCAmelCase__ = MaskGenerationPipeline(model=lowerCamelCase__ ,image_processor=lowerCamelCase__ ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : str ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def __lowerCAmelCase ( self : Optional[Any] ): pass @slow @require_torch def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = pipeline('mask-generation' ,model='facebook/sam-vit-huge' ) UpperCAmelCase__ = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] ,) # fmt: on @require_torch @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'facebook/sam-vit-huge' UpperCAmelCase__ = pipeline('mask-generation' ,model=lowerCamelCase__ ) UpperCAmelCase__ = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' ,pred_iou_thresh=1 ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] ,)
98
0
"""simple docstring""" import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy A_ = logging.get_logger(__name__) A_ = { '''artists_file''': '''artists.json''', '''lyrics_file''': '''lyrics.json''', '''genres_file''': '''genres.json''', } A_ = { '''artists_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json''', }, '''genres_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json''', }, '''lyrics_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json''', }, } A_ = { '''jukebox''': 5_12, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_LYRIC_TOKENS_SIZES lowercase__ = ["input_ids", "attention_mask"] def __init__( self: Any, a_: Dict, a_: List[str], a_: int, a_: Tuple=["v3", "v2", "v2"], a_: Tuple=512, a_: Dict=5, a_: Dict="<|endoftext|>", **a_: Tuple, ): '''simple docstring''' _snake_case : Optional[Any] = AddedToken(a_, lstrip=a_, rstrip=a_ ) if isinstance(a_, a_ ) else unk_token super().__init__( unk_token=a_, n_genres=a_, version=a_, max_n_lyric_tokens=a_, **a_, ) _snake_case : Tuple = version _snake_case : Tuple = max_n_lyric_tokens _snake_case : Optional[int] = n_genres with open(a_, encoding="""utf-8""" ) as vocab_handle: _snake_case : str = json.load(a_ ) with open(a_, encoding="""utf-8""" ) as vocab_handle: _snake_case : str = json.load(a_ ) with open(a_, encoding="""utf-8""" ) as vocab_handle: _snake_case : Dict = json.load(a_ ) _snake_case : Tuple = r"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder ) == 79: _snake_case : Dict = oov.replace(r"""\-'""", r"""\-+'""" ) _snake_case : str = regex.compile(a_ ) _snake_case : Optional[Any] = {v: k for k, v in self.artists_encoder.items()} _snake_case : Dict = {v: k for k, v in self.genres_encoder.items()} _snake_case : List[Any] = {v: k for k, v in self.lyrics_encoder.items()} @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return len(self.artists_encoder ) + len(self.genres_encoder ) + len(self.lyrics_encoder ) def UpperCamelCase_ ( self: str ): '''simple docstring''' return dict(self.artists_encoder, self.genres_encoder, self.lyrics_encoder ) def UpperCamelCase_ ( self: Any, a_: Optional[Any], a_: Optional[int], a_: Dict ): '''simple docstring''' _snake_case : Optional[Any] = [self.artists_encoder.get(a_, 0 ) for artist in list_artists] for genres in range(len(a_ ) ): _snake_case : Union[str, Any] = [self.genres_encoder.get(a_, 0 ) for genre in list_genres[genres]] _snake_case : List[Any] = list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres] )) _snake_case : Any = [[self.lyrics_encoder.get(a_, 0 ) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def UpperCamelCase_ ( self: Dict, a_: Union[str, Any] ): '''simple docstring''' return list(a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Any, a_: Optional[Any], a_: List[Any], **a_: int ): '''simple docstring''' _snake_case : Dict = self.prepare_for_tokenization(a_, a_, a_ ) _snake_case : Optional[Any] = self._tokenize(a_ ) return artist, genre, lyrics def UpperCamelCase_ ( self: str, a_: str, a_: str, a_: str, a_: bool = False ): '''simple docstring''' for idx in range(len(self.version ) ): if self.version[idx] == "v3": _snake_case : Union[str, Any] = artists[idx].lower() _snake_case : Any = [genres[idx].lower()] else: _snake_case : str = self._normalize(artists[idx] ) + """.v2""" _snake_case : List[Any] = [ self._normalize(a_ ) + """.v2""" for genre in genres[idx].split("""_""" ) ] # split is for the full dictionary with combined genres if self.version[0] == "v2": _snake_case : Tuple = regex.compile(r"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" ) _snake_case : Optional[Any] = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+'\"()[] \t\n""" _snake_case : List[Any] = {vocab[index]: index + 1 for index in range(len(a_ ) )} _snake_case : int = 0 _snake_case : List[str] = len(a_ ) + 1 _snake_case : Any = self.vocab _snake_case : List[str] = {v: k for k, v in self.vocab.items()} _snake_case : Optional[Any] = """""" else: _snake_case : Optional[int] = regex.compile(r"""[^A-Za-z0-9.,:;!?\-+'\"()\[\] \t\n]+""" ) _snake_case : List[Any] = self._run_strip_accents(a_ ) _snake_case : int = lyrics.replace("""\\""", """\n""" ) _snake_case : Optional[int] = self.out_of_vocab.sub("""""", a_ ), [], [] return artists, genres, lyrics def UpperCamelCase_ ( self: Tuple, a_: List[str] ): '''simple docstring''' _snake_case : Tuple = unicodedata.normalize("""NFD""", a_ ) _snake_case : List[Any] = [] for char in text: _snake_case : Union[str, Any] = unicodedata.category(a_ ) if cat == "Mn": continue output.append(a_ ) return "".join(a_ ) def UpperCamelCase_ ( self: Any, a_: str ): '''simple docstring''' _snake_case : Tuple = ( [chr(a_ ) for i in range(ord("""a""" ), ord("""z""" ) + 1 )] + [chr(a_ ) for i in range(ord("""A""" ), ord("""Z""" ) + 1 )] + [chr(a_ ) for i in range(ord("""0""" ), ord("""9""" ) + 1 )] + ["""."""] ) _snake_case : Union[str, Any] = frozenset(a_ ) _snake_case : Optional[Any] = re.compile(r"""_+""" ) _snake_case : List[str] = """""".join([c if c in accepted else """_""" for c in text.lower()] ) _snake_case : Optional[Any] = pattern.sub("""_""", a_ ).strip("""_""" ) return text def UpperCamelCase_ ( self: int, a_: List[str] ): '''simple docstring''' return " ".join(a_ ) def UpperCamelCase_ ( self: int, a_: Union[str, Any], a_: Optional[Union[str, TensorType]] = None, a_: bool = False ): '''simple docstring''' if not isinstance(a_, a_ ): _snake_case : int = TensorType(a_ ) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( """Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.""" ) import tensorflow as tf _snake_case : str = tf.constant _snake_case : Dict = tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError("""Unable to convert output to PyTorch tensors format, PyTorch is not installed.""" ) import torch _snake_case : Any = torch.tensor _snake_case : str = torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError("""Unable to convert output to JAX tensors format, JAX is not installed.""" ) import jax.numpy as jnp # noqa: F811 _snake_case : Any = jnp.array _snake_case : Optional[int] = _is_jax else: _snake_case : Optional[Any] = np.asarray _snake_case : Union[str, Any] = _is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: _snake_case : List[str] = [inputs] if not is_tensor(a_ ): _snake_case : Tuple = as_tensor(a_ ) except: # noqa E722 raise ValueError( """Unable to create tensor, you should probably activate truncation and/or padding """ """with 'padding=True' 'truncation=True' to have batched tensors with the same length.""" ) return inputs def __call__( self: Union[str, Any], a_: Any, a_: Tuple, a_: Dict="", a_: Dict="pt" ): '''simple docstring''' _snake_case : List[Any] = [0, 0, 0] _snake_case : Tuple = [artist] * len(self.version ) _snake_case : List[Any] = [genres] * len(self.version ) _snake_case : Dict = self.tokenize(a_, a_, a_ ) _snake_case : Dict = self._convert_token_to_id(a_, a_, a_ ) _snake_case : Union[str, Any] = [-INFINITY] * len(full_tokens[-1] ) _snake_case : Optional[Any] = [ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]], tensor_type=a_ ) for i in range(len(self.version ) ) ] return BatchEncoding({"""input_ids""": input_ids, """attention_masks""": attention_masks} ) def UpperCamelCase_ ( self: List[Any], a_: str, a_: Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a_ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _snake_case : Optional[Any] = os.path.join( a_, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""artists_file"""] ) with open(a_, """w""", encoding="""utf-8""" ) as f: f.write(json.dumps(self.artists_encoder, ensure_ascii=a_ ) ) _snake_case : Optional[Any] = os.path.join( a_, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""genres_file"""] ) with open(a_, """w""", encoding="""utf-8""" ) as f: f.write(json.dumps(self.genres_encoder, ensure_ascii=a_ ) ) _snake_case : Dict = os.path.join( a_, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""lyrics_file"""] ) with open(a_, """w""", encoding="""utf-8""" ) as f: f.write(json.dumps(self.lyrics_encoder, ensure_ascii=a_ ) ) return (artists_file, genres_file, lyrics_file) def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: List[str], a_: Tuple ): '''simple docstring''' _snake_case : str = self.artists_decoder.get(a_ ) _snake_case : Dict = [self.genres_decoder.get(a_ ) for genre in genres_index] _snake_case : Dict = [self.lyrics_decoder.get(a_ ) for character in lyric_index] return artist, genres, lyrics
354
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A_ = { '''configuration_maskformer''': ['''MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MaskFormerConfig'''], '''configuration_maskformer_swin''': ['''MaskFormerSwinConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''MaskFormerFeatureExtractor'''] A_ = ['''MaskFormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MaskFormerForInstanceSegmentation''', '''MaskFormerModel''', '''MaskFormerPreTrainedModel''', ] A_ = [ '''MaskFormerSwinBackbone''', '''MaskFormerSwinModel''', '''MaskFormerSwinPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
132
0
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self : Tuple ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = { "task_specific_params": { "summarization": {"length_penalty": 1.0, "max_length": 128, "min_length": 12, "num_beams": 4}, "summarization_cnn": {"length_penalty": 2.0, "max_length": 142, "min_length": 56, "num_beams": 4}, "summarization_xsum": {"length_penalty": 1.0, "max_length": 62, "min_length": 11, "num_beams": 6}, } } SCREAMING_SNAKE_CASE : int = { "task_specific_params.summarization.length_penalty": 1.0, "task_specific_params.summarization.max_length": 128, "task_specific_params.summarization.min_length": 12, "task_specific_params.summarization.num_beams": 4, "task_specific_params.summarization_cnn.length_penalty": 2.0, "task_specific_params.summarization_cnn.max_length": 142, "task_specific_params.summarization_cnn.min_length": 56, "task_specific_params.summarization_cnn.num_beams": 4, "task_specific_params.summarization_xsum.length_penalty": 1.0, "task_specific_params.summarization_xsum.max_length": 62, "task_specific_params.summarization_xsum.min_length": 11, "task_specific_params.summarization_xsum.num_beams": 6, } self.assertEqual(flatten_dict(a ) , a ) def __UpperCamelCase ( self : int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(a ) , x.transpose() ) ) SCREAMING_SNAKE_CASE : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def __UpperCamelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Optional[Any] = torch.tensor(a ) self.assertTrue(np.allclose(transpose(a ) , transpose(a ).numpy() ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : Optional[int] = torch.tensor(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , transpose(a , axes=(1, 2, 0) ).numpy() ) ) @require_tf def __UpperCamelCase ( self : Tuple ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Tuple = tf.constant(a ) self.assertTrue(np.allclose(transpose(a ) , transpose(a ).numpy() ) ) SCREAMING_SNAKE_CASE : int = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : Dict = tf.constant(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , transpose(a , axes=(1, 2, 0) ).numpy() ) ) @require_flax def __UpperCamelCase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : List[Any] = jnp.array(a ) self.assertTrue(np.allclose(transpose(a ) , np.asarray(transpose(a ) ) ) ) SCREAMING_SNAKE_CASE : List[Any] = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : Any = jnp.array(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , np.asarray(transpose(a , axes=(1, 2, 0) ) ) ) ) def __UpperCamelCase ( self : List[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , np.reshape(a , (4, 3) ) ) ) SCREAMING_SNAKE_CASE : Tuple = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , np.reshape(a , (12, 5) ) ) ) @require_torch def __UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : List[str] = torch.tensor(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , reshape(a , (4, 3) ).numpy() ) ) SCREAMING_SNAKE_CASE : Dict = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : Any = torch.tensor(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , reshape(a , (12, 5) ).numpy() ) ) @require_tf def __UpperCamelCase ( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Optional[int] = tf.constant(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , reshape(a , (4, 3) ).numpy() ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : int = tf.constant(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , reshape(a , (12, 5) ).numpy() ) ) @require_flax def __UpperCamelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Optional[Any] = jnp.array(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , np.asarray(reshape(a , (4, 3) ) ) ) ) SCREAMING_SNAKE_CASE : str = np.random.randn(3 , 4 , 5 ) SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.array(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , np.asarray(reshape(a , (12, 5) ) ) ) ) def __UpperCamelCase ( self : int ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(a ) , np.squeeze(a ) ) ) SCREAMING_SNAKE_CASE : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , np.squeeze(a , axis=2 ) ) ) @require_torch def __UpperCamelCase ( self : List[str] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : str = np.random.randn(1 , 3 , 4 ) SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor(a ) self.assertTrue(np.allclose(squeeze(a ) , squeeze(a ).numpy() ) ) SCREAMING_SNAKE_CASE : int = np.random.randn(1 , 4 , 1 , 5 ) SCREAMING_SNAKE_CASE : Dict = torch.tensor(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , squeeze(a , axis=2 ).numpy() ) ) @require_tf def __UpperCamelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = np.random.randn(1 , 3 , 4 ) SCREAMING_SNAKE_CASE : str = tf.constant(a ) self.assertTrue(np.allclose(squeeze(a ) , squeeze(a ).numpy() ) ) SCREAMING_SNAKE_CASE : List[str] = np.random.randn(1 , 4 , 1 , 5 ) SCREAMING_SNAKE_CASE : Any = tf.constant(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , squeeze(a , axis=2 ).numpy() ) ) @require_flax def __UpperCamelCase ( self : int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = np.random.randn(1 , 3 , 4 ) SCREAMING_SNAKE_CASE : List[str] = jnp.array(a ) self.assertTrue(np.allclose(squeeze(a ) , np.asarray(squeeze(a ) ) ) ) SCREAMING_SNAKE_CASE : int = np.random.randn(1 , 4 , 1 , 5 ) SCREAMING_SNAKE_CASE : str = jnp.array(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , np.asarray(squeeze(a , axis=2 ) ) ) ) def __UpperCamelCase ( self : str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , np.expand_dims(a , axis=1 ) ) ) @require_torch def __UpperCamelCase ( self : Optional[int] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : str = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Tuple = torch.tensor(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , expand_dims(a , axis=1 ).numpy() ) ) @require_tf def __UpperCamelCase ( self : Optional[int] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Dict = tf.constant(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , expand_dims(a , axis=1 ).numpy() ) ) @require_flax def __UpperCamelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = np.random.randn(3 , 4 ) SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.array(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , np.asarray(expand_dims(a , axis=1 ) ) ) )
76
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES A_ : str = logging.get_logger(__name__) A_ : str = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) A_ : Optional[int] = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) A_ : Union[str, Any] = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) A_ : Dict = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) A_ : Optional[int] = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) A_ : Dict = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) A_ : List[str] = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) A_ : Tuple = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) A_ : Optional[int] = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) A_ : int = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) A_ : Tuple = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) A_ : Tuple = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) A_ : int = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) A_ : Tuple = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) A_ : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) A_ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) A_ : Dict = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) A_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) A_ : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) A_ : Dict = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) A_ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) A_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) A_ : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) A_ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) A_ : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) A_ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) A_ : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) A_ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_MAPPING A_ : Tuple = auto_class_update(FlaxAutoModel) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_PRETRAINING_MAPPING A_ : str = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING A_ : Optional[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_MASKED_LM_MAPPING A_ : List[str] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING A_ : Union[str, Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING A_ : Tuple = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING A_ : Any = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ : Dict = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING A_ : Any = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING A_ : Tuple = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING A_ : int = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING A_ : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class A_ ( _BaseAutoModelClass ): '''simple docstring''' a__ = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING A_ : Optional[int] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
333
0
"""simple docstring""" def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase ) -> int: '''simple docstring''' while second != 0: lowercase : Any = first & second first ^= second lowercase : Any = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() _UpperCamelCase: str = int(input('Enter the first number: ').strip()) _UpperCamelCase: Any = int(input('Enter the second number: ').strip()) print(f'''{add(first, second) = }''')
53
"""simple docstring""" from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _UpperCamelCase: Dict = logging.get_logger(__name__) def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: '''simple docstring''' return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None ) -> str: '''simple docstring''' lowercase : str = tesseract_config if tesseract_config is not None else '' # apply OCR lowercase : Tuple = to_pil_image(_UpperCAmelCase ) lowercase , lowercase : Union[str, Any] = pil_image.size lowercase : Union[str, Any] = pytesseract.image_to_data(_UpperCAmelCase , lang=_UpperCAmelCase , output_type='dict' , config=_UpperCAmelCase ) lowercase , lowercase , lowercase , lowercase , lowercase : str = data['text'], data['left'], data['top'], data['width'], data['height'] # filter empty words and corresponding coordinates lowercase : str = [idx for idx, word in enumerate(_UpperCAmelCase ) if not word.strip()] lowercase : Dict = [word for idx, word in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase : Union[str, Any] = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase : str = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase : Any = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase : Dict = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format lowercase : int = [] for x, y, w, h in zip(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowercase : Any = [x, y, x + w, y + h] actual_boxes.append(_UpperCAmelCase ) # finally, normalize the bounding boxes lowercase : List[str] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) ) assert len(_UpperCAmelCase ) == len(_UpperCAmelCase ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a__ ( SCREAMING_SNAKE_CASE__ ): _lowerCamelCase = ['pixel_values'] def __init__( self : str, lowerCAmelCase : bool = True, lowerCAmelCase : Dict[str, int] = None, lowerCAmelCase : PILImageResampling = PILImageResampling.BILINEAR, lowerCAmelCase : bool = True, lowerCAmelCase : Optional[str] = None, lowerCAmelCase : Optional[str] = "", **lowerCAmelCase : List[Any], ) -> None: super().__init__(**lowerCAmelCase ) lowercase : Optional[Any] = size if size is not None else {'height': 224, 'width': 224} lowercase : List[Any] = get_size_dict(lowerCAmelCase ) lowercase : str = do_resize lowercase : List[str] = size lowercase : int = resample lowercase : List[str] = apply_ocr lowercase : str = ocr_lang lowercase : Union[str, Any] = tesseract_config def lowercase ( self : Optional[Any], lowerCAmelCase : np.ndarray, lowerCAmelCase : Dict[str, int], lowerCAmelCase : PILImageResampling = PILImageResampling.BILINEAR, lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None, **lowerCAmelCase : int, ) -> np.ndarray: lowercase : Optional[Any] = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) lowercase : Optional[int] = (size['height'], size['width']) return resize(lowerCAmelCase, size=lowerCAmelCase, resample=lowerCAmelCase, data_format=lowerCAmelCase, **lowerCAmelCase ) def lowercase ( self : Any, lowerCAmelCase : ImageInput, lowerCAmelCase : bool = None, lowerCAmelCase : Dict[str, int] = None, lowerCAmelCase : PILImageResampling = None, lowerCAmelCase : bool = None, lowerCAmelCase : Optional[str] = None, lowerCAmelCase : Optional[str] = None, lowerCAmelCase : Optional[Union[str, TensorType]] = None, lowerCAmelCase : ChannelDimension = ChannelDimension.FIRST, **lowerCAmelCase : List[Any], ) -> PIL.Image.Image: lowercase : Any = do_resize if do_resize is not None else self.do_resize lowercase : Union[str, Any] = size if size is not None else self.size lowercase : Dict = get_size_dict(lowerCAmelCase ) lowercase : List[str] = resample if resample is not None else self.resample lowercase : str = apply_ocr if apply_ocr is not None else self.apply_ocr lowercase : Tuple = ocr_lang if ocr_lang is not None else self.ocr_lang lowercase : List[str] = tesseract_config if tesseract_config is not None else self.tesseract_config lowercase : Optional[int] = make_list_of_images(lowerCAmelCase ) if not valid_images(lowerCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) # All transformations expect numpy arrays. lowercase : int = [to_numpy_array(lowerCAmelCase ) for image in images] if apply_ocr: requires_backends(self, 'pytesseract' ) lowercase : str = [] lowercase : Dict = [] for image in images: lowercase , lowercase : List[str] = apply_tesseract(lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ) words_batch.append(lowerCAmelCase ) boxes_batch.append(lowerCAmelCase ) if do_resize: lowercase : str = [self.resize(image=lowerCAmelCase, size=lowerCAmelCase, resample=lowerCAmelCase ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) lowercase : Any = [flip_channel_order(lowerCAmelCase ) for image in images] lowercase : Dict = [to_channel_dimension_format(lowerCAmelCase, lowerCAmelCase ) for image in images] lowercase : Tuple = BatchFeature(data={'pixel_values': images}, tensor_type=lowerCAmelCase ) if apply_ocr: lowercase : List[Any] = words_batch lowercase : Tuple = boxes_batch return data
53
1
"""simple docstring""" def A ( snake_case :list , snake_case :list , snake_case :int ) -> int: if len(snake_case ) != len(snake_case ): raise ValueError('The length of profit and weight must be same.' ) if max_weight <= 0: raise ValueError('max_weight must greater than zero.' ) if any(p < 0 for p in profit ): raise ValueError('Profit can not be negative.' ) if any(w < 0 for w in weight ): raise ValueError('Weight can not be negative.' ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. __UpperCamelCase = [p / w for p, w in zip(snake_case , snake_case )] # Creating a copy of the list and sorting profit/weight in ascending order __UpperCamelCase = sorted(snake_case ) # declaring useful variables __UpperCamelCase = len(snake_case ) __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight __UpperCamelCase = sorted_profit_by_weight[length - i - 1] __UpperCamelCase = profit_by_weight.index(snake_case ) __UpperCamelCase = -1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( "Input profits, weights, and then max_weight (all positive ints) separated by " "spaces." ) UpperCamelCase : str = [int(x) for x in input("Input profits separated by spaces: ").split()] UpperCamelCase : str = [int(x) for x in input("Input weights separated by spaces: ").split()] UpperCamelCase : str = int(input("Max weight allowed: ")) # Function Call calc_profit(profit, weight, max_weight)
316
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys UpperCamelCase : Union[str, Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") UpperCamelCase : Any = subprocess.check_output(f'''git diff --name-only {fork_point_sha}'''.split()).decode("utf-8").split() UpperCamelCase : Tuple = "|".join(sys.argv[1:]) UpperCamelCase : Optional[int] = re.compile(Rf'''^({joined_dirs}).*?\.py$''') UpperCamelCase : Optional[Any] = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
316
1
snake_case_ : int = {str(digit): digit**5 for digit in range(10)} def A (__A : int ) -> int: """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(__A ) ) def A () -> int: """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(__A ) ) if __name__ == "__main__": print(solution())
7
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class __snake_case ( unittest.TestCase ): def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = 10 def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ = [1, 2, 3, 4] UpperCAmelCase_ = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(_snake_case , self.block_size , 0) , _snake_case) def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] UpperCAmelCase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_snake_case , self.block_size , 0) , _snake_case) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] UpperCAmelCase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_snake_case , self.block_size , 0) , _snake_case) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = '''It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.''' UpperCAmelCase_ , UpperCAmelCase_ = process_story(_snake_case) self.assertEqual(_snake_case , []) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = '''''' UpperCAmelCase_ , UpperCAmelCase_ = process_story(_snake_case) self.assertEqual(_snake_case , []) self.assertEqual(_snake_case , []) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = ( '''It was the year of Our Lord one thousand seven hundred and ''' '''seventy-five\n\nSpiritual revelations were conceded to England ''' '''at that favoured period, as at this.\n@highlight\n\nIt was the best of times''' ) UpperCAmelCase_ , UpperCAmelCase_ = process_story(_snake_case) UpperCAmelCase_ = [ '''It was the year of Our Lord one thousand seven hundred and seventy-five.''', '''Spiritual revelations were conceded to England at that favoured period, as at this.''', ] self.assertEqual(_snake_case , _snake_case) UpperCAmelCase_ = ['''It was the best of times.'''] self.assertEqual(_snake_case , _snake_case) def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = torch.tensor([1, 2, 3, 4]) UpperCAmelCase_ = torch.tensor([1, 1, 1, 1]) np.testing.assert_array_equal(build_mask(_snake_case , 0).numpy() , expected.numpy()) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = torch.tensor([1, 2, 3, 4, 23, 23, 23]) UpperCAmelCase_ = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(_snake_case , 23).numpy() , expected.numpy()) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = torch.tensor([8, 2, 3, 4, 1, 1, 1]) UpperCAmelCase_ = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(_snake_case , 1).numpy() , expected.numpy()) def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = 101 UpperCAmelCase_ = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]]) UpperCAmelCase_ = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]]) UpperCAmelCase_ = compute_token_type_ids(_snake_case , _snake_case) np.testing.assert_array_equal(_snake_case , _snake_case)
7
1
import os import jsonlines import numpy as np from tqdm import tqdm _UpperCamelCase = 2048 _UpperCamelCase = 4096 _UpperCamelCase = 42 _UpperCamelCase = os.environ.pop("PROCESS_TRAIN", "false") _UpperCamelCase = {"null": 0, "short": 1, "long": 2, "yes": 3, "no": 4} def _lowercase ( lowercase__ ): def choose_first(lowercase__ , lowercase__=False ): assert isinstance(lowercase__ , lowercase__ ) if len(lowercase__ ) == 1: __lowerCAmelCase : Tuple = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: __lowerCAmelCase : str = {k: [a[k]] for k in a} if len(a['''start_token'''] ) > 0: break return a __lowerCAmelCase : Optional[Any] = {'''id''': example['''id''']} __lowerCAmelCase : List[str] = example['''annotations'''] __lowerCAmelCase : Optional[int] = annotation['''yes_no_answer'''] if 0 in yes_no_answer or 1 in yes_no_answer: __lowerCAmelCase : int = ['''yes'''] if 1 in yes_no_answer else ['''no'''] __lowerCAmelCase : Optional[int] = [] __lowerCAmelCase : Any = [] __lowerCAmelCase : List[str] = ['''<cls>'''] else: __lowerCAmelCase : Any = ['''short'''] __lowerCAmelCase : List[str] = choose_first(annotation['''short_answers'''] ) if len(out['''start_token'''] ) == 0: # answer will be long if short is not available __lowerCAmelCase : int = ['''long'''] __lowerCAmelCase : int = choose_first(annotation['''long_answer'''] , is_long_answer=lowercase__ ) __lowerCAmelCase : Dict = [] answer.update(lowercase__ ) # disregard some samples if len(answer['''start_token'''] ) > 1 or answer["start_token"] == answer["end_token"]: __lowerCAmelCase : Optional[Any] = True else: __lowerCAmelCase : str = False __lowerCAmelCase : str = ['''start_token''', '''end_token''', '''start_byte''', '''end_byte''', '''text'''] if not all(isinstance(answer[k] , lowercase__ ) for k in cols ): raise ValueError('''Issue in ID''' , example['''id'''] ) return answer def _lowercase ( lowercase__ , lowercase__=False ): __lowerCAmelCase : Dict = _get_single_answer(lowercase__ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element __lowerCAmelCase : Optional[Any] = example['''document''']['''tokens'''] __lowerCAmelCase : Union[str, Any] = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) return { "context": " ".join(lowercase__ ), "answer": { "start_token": -1_0_0, # ignore index in cross-entropy "end_token": -1_0_0, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples __lowerCAmelCase : Any = ['''start_token''', '''end_token'''] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 __lowerCAmelCase : Optional[Any] = example['''document''']['''tokens'''] __lowerCAmelCase : List[Any] = answer['''start_token'''] __lowerCAmelCase : str = answer['''end_token'''] __lowerCAmelCase : Dict = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 __lowerCAmelCase : Optional[Any] = ''' '''.join(context[start_token:end_token] ) # checking above code if assertion: __lowerCAmelCase : List[Any] = doc['''is_html'''][answer['''start_token'''] : answer['''end_token''']] __lowerCAmelCase : List[str] = doc['''token'''][answer['''start_token'''] : answer['''end_token''']] __lowerCAmelCase : Optional[Any] = ''' '''.join([old[i] for i in range(len(lowercase__ ) ) if not is_html[i]] ) if new != old: print('''ID:''' , example['''id'''] ) print('''New:''' , lowercase__ , end='''\n''' ) print('''Old:''' , lowercase__ , end='''\n\n''' ) return { "context": " ".join(lowercase__ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def _lowercase ( lowercase__ , lowercase__ , lowercase__=2_0_4_8 , lowercase__=4_0_9_6 , lowercase__=True ): # overlap will be of doc_stride - q_len __lowerCAmelCase : int = get_context_and_ans(lowercase__ , assertion=lowercase__ ) __lowerCAmelCase : str = out['''answer'''] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } __lowerCAmelCase : List[Any] = tokenizer(example['''question''']['''text'''] , out['''context'''] ).input_ids __lowerCAmelCase : List[Any] = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Any = [] __lowerCAmelCase : Any = input_ids[:q_len] __lowerCAmelCase : str = range(lowercase__ , len(lowercase__ ) , max_length - doc_stride ) for i in doc_start_indices: __lowerCAmelCase : Optional[Any] = i + max_length - q_len __lowerCAmelCase : Dict = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['''category'''][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-1_0_0] * len(lowercase__ ), "end_token": [-1_0_0] * len(lowercase__ ), "category": category, }, } __lowerCAmelCase : Tuple = out['''context'''].split() __lowerCAmelCase : Tuple = splitted_context[answer['''end_token''']] __lowerCAmelCase : Optional[int] = len( tokenizer( ''' '''.join(splitted_context[: answer['''start_token''']] ) , add_special_tokens=lowercase__ , ).input_ids ) __lowerCAmelCase : str = len( tokenizer(''' '''.join(splitted_context[: answer['''end_token''']] ) , add_special_tokens=lowercase__ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token __lowerCAmelCase : Any = len(tokenizer(lowercase__ , add_special_tokens=lowercase__ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 __lowerCAmelCase : Dict = input_ids[answer['''start_token'''] : answer['''end_token'''] + 1] # right & left are inclusive __lowerCAmelCase : Optional[int] = answer['''start_token'''] __lowerCAmelCase : str = answer['''end_token'''] if assertion: __lowerCAmelCase : Optional[int] = tokenizer.decode(lowercase__ ) if answer["span"] != new: print('''ISSUE IN TOKENIZATION''' ) print('''OLD:''' , answer['''span'''] ) print('''NEW:''' , lowercase__ , end='''\n\n''' ) if len(lowercase__ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } __lowerCAmelCase : int = input_ids[:q_len] __lowerCAmelCase : Tuple = range(lowercase__ , len(lowercase__ ) , max_length - doc_stride ) __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Dict = [] __lowerCAmelCase : int = [] __lowerCAmelCase : Optional[Any] = [] # null, yes, no, long, short for i in doc_start_indices: __lowerCAmelCase : int = i + max_length - q_len __lowerCAmelCase : Tuple = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: __lowerCAmelCase : List[Any] = start_token - i + q_len __lowerCAmelCase : Union[str, Any] = end_token - i + q_len answers_category.append(answer['''category'''][0] ) # ["short"] -> "short" else: __lowerCAmelCase : Any = -1_0_0 __lowerCAmelCase : Tuple = -1_0_0 answers_category.append('''null''' ) __lowerCAmelCase : int = inputs[-1][start_token : end_token + 1] answers_start_token.append(lowercase__ ) answers_end_token.append(lowercase__ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('''ISSUE in strided for ID:''' , example['''id'''] ) print('''New:''' , tokenizer.decode(lowercase__ ) ) print('''Old:''' , tokenizer.decode(lowercase__ ) , end='''\n\n''' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def _lowercase ( lowercase__ , lowercase__ , lowercase__=2_0_4_8 , lowercase__=4_0_9_6 , lowercase__=False ): __lowerCAmelCase : List[Any] = get_strided_contexts_and_ans( lowercase__ , lowercase__ , doc_stride=lowercase__ , max_length=lowercase__ , assertion=lowercase__ , ) return example def _lowercase ( lowercase__ , lowercase__ ): with jsonlines.open(lowercase__ , '''a''' ) as writer: for example in tqdm(lowercase__ , total=len(lowercase__ ) , desc='''Saving samples ... ''' ): __lowerCAmelCase : Tuple = example['''labels'''] for ids, start, end, cat in zip( example['''input_ids'''] , labels['''start_token'''] , labels['''end_token'''] , labels['''category'''] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { '''input_ids''': ids, '''start_token''': start, '''end_token''': end, '''category''': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer _UpperCamelCase = load_dataset("natural_questions") _UpperCamelCase = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base") _UpperCamelCase = data["train" if PROCESS_TRAIN == "true" else "validation"] _UpperCamelCase = { "tokenizer": tokenizer, "doc_stride": DOC_STRIDE, "max_length": MAX_LENGTH, "assertion": False, } _UpperCamelCase = data.map(prepare_inputs, fn_kwargs=fn_kwargs) _UpperCamelCase = data.remove_columns(["annotations", "document", "id", "question"]) print(data) np.random.seed(SEED) _UpperCamelCase = "nq-training.jsonl" if PROCESS_TRAIN == "true" else "nq-validation.jsonl" save_to_disk(data, file_name=cache_file_name)
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1
'''simple docstring''' from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = int(number**0.5 ) return number == sq * sq def _lowerCAmelCase ( lowerCamelCase_ : int , lowerCamelCase_ : int , lowerCamelCase_ : int , lowerCamelCase_ : int , lowerCamelCase_ : int , lowerCamelCase_ : int ): __lowercase = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den __lowercase = x_den * y_den * z_den __lowercase = gcd(lowerCamelCase_ , lowerCamelCase_ ) top //= hcf bottom //= hcf return top, bottom def _lowerCAmelCase ( lowerCamelCase_ : int = 3_5 ): __lowercase = set() __lowercase = 42 __lowercase = Fraction(0 ) __lowercase = 42 for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 __lowercase = x_num * y_den + x_den * y_num __lowercase = x_den * y_den __lowercase = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: __lowercase = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 __lowercase = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) __lowercase = x_den * x_den * y_den * y_den if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): __lowercase = int(sqrt(lowerCamelCase_ ) ) __lowercase = int(sqrt(lowerCamelCase_ ) ) __lowercase = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: __lowercase = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=-1 __lowercase = x_num * y_num __lowercase = x_den * y_num + x_num * y_den __lowercase = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: __lowercase = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 __lowercase = x_num * x_num * y_num * y_num __lowercase = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): __lowercase = int(sqrt(lowerCamelCase_ ) ) __lowercase = int(sqrt(lowerCamelCase_ ) ) __lowercase = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: __lowercase = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) for num, den in unique_s: total += Fraction(lowerCamelCase_ , lowerCamelCase_ ) return total.denominator + total.numerator if __name__ == "__main__": print(f'''{solution() = }''')
217
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { '''Intel/dpt-large''': '''https://huggingface.co/Intel/dpt-large/resolve/main/config.json''', # See all DPT models at https://huggingface.co/models?filter=dpt } class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' a : Union[str, Any] = "dpt" def __init__(self ,_lowerCamelCase=768 ,_lowerCamelCase=12 ,_lowerCamelCase=12 ,_lowerCamelCase=3072 ,_lowerCamelCase="gelu" ,_lowerCamelCase=0.0 ,_lowerCamelCase=0.0 ,_lowerCamelCase=0.0_2 ,_lowerCamelCase=1E-1_2 ,_lowerCamelCase=384 ,_lowerCamelCase=16 ,_lowerCamelCase=3 ,_lowerCamelCase=False ,_lowerCamelCase=True ,_lowerCamelCase=[2, 5, 8, 11] ,_lowerCamelCase="project" ,_lowerCamelCase=[4, 2, 1, 0.5] ,_lowerCamelCase=[96, 192, 384, 768] ,_lowerCamelCase=256 ,_lowerCamelCase=-1 ,_lowerCamelCase=False ,_lowerCamelCase=True ,_lowerCamelCase=0.4 ,_lowerCamelCase=255 ,_lowerCamelCase=0.1 ,_lowerCamelCase=[1, 1024, 24, 24] ,_lowerCamelCase=[0, 1] ,_lowerCamelCase=None ,**_lowerCamelCase ,) -> Tuple: '''simple docstring''' super().__init__(**_lowerCamelCase ) __lowercase = hidden_size __lowercase = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info('''Initializing the config with a `BiT` backbone.''' ) __lowercase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, } __lowercase = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase ,_lowerCamelCase ): logger.info('''Initializing the config with a `BiT` backbone.''' ) __lowercase = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase ,_lowerCamelCase ): __lowercase = backbone_config else: raise ValueError( f"backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}." ) __lowercase = backbone_featmap_shape __lowercase = neck_ignore_stages if readout_type != "project": raise ValueError('''Readout type must be \'project\' when using `DPT-hybrid` mode.''' ) else: __lowercase = None __lowercase = None __lowercase = [] __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError('''Readout_type must be one of [\'ignore\', \'add\', \'project\']''' ) __lowercase = readout_type __lowercase = reassemble_factors __lowercase = neck_hidden_sizes __lowercase = fusion_hidden_size __lowercase = head_in_index __lowercase = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) __lowercase = use_auxiliary_head __lowercase = auxiliary_loss_weight __lowercase = semantic_loss_ignore_index __lowercase = semantic_classifier_dropout def _UpperCAmelCase (self ) -> Union[str, Any]: '''simple docstring''' __lowercase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output
217
1
import mpmath # for roots of unity import numpy as np class UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[int] , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : str=None ): """simple docstring""" _A: Optional[int] = list(poly_a or [0] )[:] _A: List[str] = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() _A: Any = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() _A: Optional[int] = len(self.polyB ) # Add 0 to make lengths equal a power of 2 _A: Dict = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform _A: List[str] = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product _A: Tuple = self.__multiply() def __magic_name__ ( self : int , lowerCAmelCase_ : List[Any] ): """simple docstring""" _A: Union[str, Any] = [[x] for x in self.polyA] if which == """A""" else [[x] for x in self.polyB] # Corner case if len(lowerCAmelCase_ ) <= 1: return dft[0] # _A: Union[str, Any] = self.c_max_length // 2 while next_ncol > 0: _A: Optional[Any] = [[] for i in range(lowerCAmelCase_ )] _A: List[Any] = self.root**next_ncol # First half of next step _A: List[str] = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(lowerCAmelCase_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step _A: int = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(lowerCAmelCase_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update _A: Dict = new_dft _A: Dict = next_ncol // 2 return dft[0] def __magic_name__ ( self : Any ): """simple docstring""" _A: str = self.__dft('''A''' ) _A: str = self.__dft('''B''' ) _A: str = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT _A: Optional[Any] = 2 while next_ncol <= self.c_max_length: _A: str = [[] for i in range(lowerCAmelCase_ )] _A: Any = self.root ** (next_ncol // 2) _A: Union[str, Any] = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update _A: Optional[Any] = new_inverse_c next_ncol *= 2 # Unpack _A: Any = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1J for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Tuple ): """simple docstring""" _A: int = """A = """ + """ + """.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.polyA[: self.len_A] ) ) _A: Optional[Any] = """B = """ + """ + """.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.polyB[: self.len_B] ) ) _A: Union[str, Any] = """A*B = """ + """ + """.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.product ) ) return F"""{a}\n{b}\n{c}""" # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
121
"""simple docstring""" import datasets from .evaluate import evaluate _UpperCAmelCase = """\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } """ _UpperCAmelCase = """ This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. """ _UpperCAmelCase = """ Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair as given in the references (see below) - 'prediction_text': the text of the answer references: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair (see above), - 'answers': a Dict in the SQuAD dataset format { 'text': list of possible texts for the answer, as a list of strings 'answer_start': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: 'exact_match': Exact match (the normalized answer exactly match the gold answer) 'f1': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}] >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}] >>> squad_metric = datasets.load_metric(\"squad\") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 100.0, 'f1': 100.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def lowerCamelCase__ ( self : int ) -> List[str]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def lowerCamelCase__ ( self : int , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Optional[int] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: Dict ={prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} SCREAMING_SNAKE_CASE_: Tuple =[ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] SCREAMING_SNAKE_CASE_: str =evaluate(dataset=lowerCAmelCase , predictions=lowerCAmelCase ) return score
173
0
'''simple docstring''' import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def a_ ( __snake_case : List[str] ) -> Dict: """simple docstring""" monkeypatch.setattr('''datasets.utils.deprecation_utils._emitted_deprecation_warnings''' , set() ) @pytest.fixture def a_ ( __snake_case : Dict ) -> Tuple: """simple docstring""" class __UpperCamelCase : def __init__( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =metric_id class __UpperCamelCase : lowercase : Union[str, Any] =[MetricMock(lowerCamelCase__ ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def lowercase__ ( self ): """simple docstring""" return self._metrics monkeypatch.setattr('''datasets.inspect.huggingface_hub''' , HfhMock() ) @pytest.mark.parametrize( '''func, args''' , [(load_metric, ('''metrics/mse''',)), (list_metrics, ()), (inspect_metric, ('''metrics/mse''', '''tmp_path'''))] ) def a_ ( __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : Any , __snake_case : Union[str, Any] , __snake_case : Optional[Any] ) -> Union[str, Any]: """simple docstring""" if "tmp_path" in args: lowerCamelCase_ =tuple(arg if arg != '''tmp_path''' else tmp_path for arg in args ) with pytest.warns(__snake_case , match='''https://huggingface.co/docs/evaluate''' ): func(*__snake_case )
6
'''simple docstring''' import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def a_ ( __snake_case : Optional[int] ) -> List[str]: """simple docstring""" lowerCamelCase_ =[ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(__snake_case , __snake_case ) def a_ ( __snake_case : List[Any] ) -> int: """simple docstring""" lowerCamelCase_, lowerCamelCase_ =emb.weight.shape lowerCamelCase_ =nn.Linear(__snake_case , __snake_case , bias=__snake_case ) lowerCamelCase_ =emb.weight.data return lin_layer def a_ ( __snake_case : Union[str, Any] , __snake_case : Tuple=None ) -> Dict: """simple docstring""" lowerCamelCase_ ={} for old_key in state_dict.keys(): lowerCamelCase_ =old_key if "moe_layer.experts." in key: if expert_idx is not None: lowerCamelCase_ =key.replace('''moe_layer.experts.0''' , F'''ffn.experts.expert_{expert_idx}''' ) else: lowerCamelCase_ =key.replace('''moe_layer.experts.''' , '''ffn.experts.expert_''' ) if "gate" in key: lowerCamelCase_ =key.replace('''.moe_layer.gate.wg''' , '''.ffn.router.classifier''' ) if "fc2" and "experts" not in key: lowerCamelCase_ =key.replace('''.fc2.''' , '''.ffn.fc2.''' ) if "fc1" and "experts" not in key: lowerCamelCase_ =key.replace('''.fc1.''' , '''.ffn.fc1.''' ) if ".encoder_attn." in key: lowerCamelCase_ =key.replace('''.encoder_attn.''' , '''.cross_attention.''' ) if "encoder_attn_layer_norm" in key: lowerCamelCase_ =key.replace('''encoder_attn_layer_norm''' , '''cross_attention_layer_norm''' ) if "final_layer_norm" in key: lowerCamelCase_ =key.replace('''final_layer_norm''' , '''ff_layer_norm''' ) lowerCamelCase_ =state_dict[old_key] return new_dict def a_ ( __snake_case : Optional[Any] , __snake_case : Any , __snake_case : Optional[int] , __snake_case : Tuple , __snake_case : str = WEIGHTS_NAME ) -> Dict: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =0 os.makedirs(__snake_case , exist_ok=__snake_case ) for expert in range(__snake_case ): lowerCamelCase_ =switch_checkpoint_path + F'''-rank-{expert}.pt''' if os.path.isfile(__snake_case ): lowerCamelCase_ =torch.load(__snake_case )['''model'''] remove_ignore_keys_(__snake_case ) lowerCamelCase_ =rename_fairseq_keys(__snake_case , __snake_case ) lowerCamelCase_ =os.path.join( __snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) torch.save(__snake_case , __snake_case ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(__snake_case )[0]].dtype ) # Add the last block lowerCamelCase_ =os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) lowerCamelCase_ =torch.load(switch_checkpoint_path + '''-shared.pt''' )['''model'''] remove_ignore_keys_(__snake_case ) lowerCamelCase_ =rename_fairseq_keys(__snake_case , __snake_case ) lowerCamelCase_ =shared_weights['''decoder.embed_tokens.weight'''] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(__snake_case ) == 1: lowerCamelCase_ =os.path.join(__snake_case , __snake_case ) torch.save(__snake_case , __snake_case ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(__snake_case , __snake_case ) # Otherwise, let's build the index lowerCamelCase_ ={} for idx, shard in enumerate(__snake_case ): lowerCamelCase_ =weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-{len(__snake_case ):05d}.bin''' ) lowerCamelCase_ =os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(__snake_case , os.path.join(__snake_case , __snake_case ) ) for key in shard: lowerCamelCase_ =shard_file # Add the metadata lowerCamelCase_ ={'''total_size''': total_size} lowerCamelCase_ ={'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(__snake_case , __snake_case ) , '''w''' , encoding='''utf-8''' ) as f: lowerCamelCase_ =json.dumps(__snake_case , indent=2 , sort_keys=__snake_case ) + '''\n''' f.write(__snake_case ) return metadata, index if __name__ == "__main__": a_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--nllb_moe_checkpoint_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--dtype""", default="""float32""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b""", type=str, required=False, help="""Path to the output pytorch model.""", ) a_ : Tuple = parser.parse_args() a_ , a_ : int = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 1_28, args.dtype, ) a_ : Tuple = NllbMoeConfig.from_pretrained( """facebook/nllb-200-3.3B""", encoder_sparse_step=4, decoder_sparse_step=4, num_experts=1_28 ) config.save_pretrained(args.pytorch_dump_folder_path) a_ : Any = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print("""Done""") model.save_pretrained(args.pytorch_dump_folder_path)
6
1
'''simple docstring''' from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class A__ ( _snake_case ): lowercase = 42 class A__ ( _snake_case , _snake_case ): lowercase = True @register_to_config def __init__( self , UpperCamelCase__ = 3 , UpperCamelCase__ = 3 , UpperCamelCase__ = ("DownEncoderBlock2D",) , UpperCamelCase__ = ("UpDecoderBlock2D",) , UpperCamelCase__ = (64,) , UpperCamelCase__ = 1 , UpperCamelCase__ = "silu" , UpperCamelCase__ = 4 , UpperCamelCase__ = 32 , UpperCamelCase__ = 32 , UpperCamelCase__ = 0.18215 , ) -> Optional[int]: '''simple docstring''' super().__init__() # pass init params to Encoder A_ = Encoder( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , down_block_types=UpperCamelCase__ , block_out_channels=UpperCamelCase__ , layers_per_block=UpperCamelCase__ , act_fn=UpperCamelCase__ , norm_num_groups=UpperCamelCase__ , double_z=UpperCamelCase__ , ) # pass init params to Decoder A_ = Decoder( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , up_block_types=UpperCamelCase__ , block_out_channels=UpperCamelCase__ , layers_per_block=UpperCamelCase__ , norm_num_groups=UpperCamelCase__ , act_fn=UpperCamelCase__ , ) A_ = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) A_ = nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A_ = False A_ = False # only relevant if vae tiling is enabled A_ = self.config.sample_size A_ = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) A_ = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) A_ = 0.25 def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__=False ) -> Any: '''simple docstring''' if isinstance(UpperCamelCase__ , (Encoder, Decoder) ): A_ = value def snake_case_ ( self , UpperCamelCase__ = True ) -> List[str]: '''simple docstring''' A_ = use_tiling def snake_case_ ( self ) -> List[Any]: '''simple docstring''' self.enable_tiling(UpperCamelCase__ ) def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = True def snake_case_ ( self ) -> Optional[Any]: '''simple docstring''' A_ = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def snake_case_ ( self ) -> Dict[str, AttentionProcessor]: '''simple docstring''' A_ = {} def fn_recursive_add_processors(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): if hasattr(UpperCamelCase__ , """set_processor""" ): A_ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , UpperCamelCase__ , UpperCamelCase__ ) return processors for name, module in self.named_children(): fn_recursive_add_processors(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return processors def snake_case_ ( self , UpperCamelCase__ ) -> Dict: '''simple docstring''' A_ = len(self.attn_processors.keys() ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(UpperCamelCase__ )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): if hasattr(UpperCamelCase__ , """set_processor""" ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): module.set_processor(UpperCamelCase__ ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , UpperCamelCase__ , UpperCamelCase__ ) for name, module in self.named_children(): fn_recursive_attn_processor(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def snake_case_ ( self ) -> Any: '''simple docstring''' self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = True ) -> AutoencoderKLOutput: '''simple docstring''' if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(UpperCamelCase__ , return_dict=UpperCamelCase__ ) if self.use_slicing and x.shape[0] > 1: A_ = [self.encoder(UpperCamelCase__ ) for x_slice in x.split(1 )] A_ = torch.cat(UpperCamelCase__ ) else: A_ = self.encoder(UpperCamelCase__ ) A_ = self.quant_conv(UpperCamelCase__ ) A_ = DiagonalGaussianDistribution(UpperCamelCase__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = True ) -> Union[DecoderOutput, torch.FloatTensor]: '''simple docstring''' if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(UpperCamelCase__ , return_dict=UpperCamelCase__ ) A_ = self.post_quant_conv(UpperCamelCase__ ) A_ = self.decoder(UpperCamelCase__ ) if not return_dict: return (dec,) return DecoderOutput(sample=UpperCamelCase__ ) @apply_forward_hook def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = True ) -> Union[DecoderOutput, torch.FloatTensor]: '''simple docstring''' if self.use_slicing and z.shape[0] > 1: A_ = [self._decode(UpperCamelCase__ ).sample for z_slice in z.split(1 )] A_ = torch.cat(UpperCamelCase__ ) else: A_ = self._decode(UpperCamelCase__ ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: '''simple docstring''' A_ = min(a.shape[2] , b.shape[2] , UpperCamelCase__ ) for y in range(UpperCamelCase__ ): A_ = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' A_ = min(a.shape[3] , b.shape[3] , UpperCamelCase__ ) for x in range(UpperCamelCase__ ): A_ = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = True ) -> AutoencoderKLOutput: '''simple docstring''' A_ = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) A_ = int(self.tile_latent_min_size * self.tile_overlap_factor ) A_ = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. A_ = [] for i in range(0 , x.shape[2] , UpperCamelCase__ ): A_ = [] for j in range(0 , x.shape[3] , UpperCamelCase__ ): A_ = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] A_ = self.encoder(UpperCamelCase__ ) A_ = self.quant_conv(UpperCamelCase__ ) row.append(UpperCamelCase__ ) rows.append(UpperCamelCase__ ) A_ = [] for i, row in enumerate(UpperCamelCase__ ): A_ = [] for j, tile in enumerate(UpperCamelCase__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: A_ = self.blend_v(rows[i - 1][j] , UpperCamelCase__ , UpperCamelCase__ ) if j > 0: A_ = self.blend_h(row[j - 1] , UpperCamelCase__ , UpperCamelCase__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(UpperCamelCase__ , dim=3 ) ) A_ = torch.cat(UpperCamelCase__ , dim=2 ) A_ = DiagonalGaussianDistribution(UpperCamelCase__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = True ) -> Union[DecoderOutput, torch.FloatTensor]: '''simple docstring''' A_ = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) A_ = int(self.tile_sample_min_size * self.tile_overlap_factor ) A_ = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. A_ = [] for i in range(0 , z.shape[2] , UpperCamelCase__ ): A_ = [] for j in range(0 , z.shape[3] , UpperCamelCase__ ): A_ = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] A_ = self.post_quant_conv(UpperCamelCase__ ) A_ = self.decoder(UpperCamelCase__ ) row.append(UpperCamelCase__ ) rows.append(UpperCamelCase__ ) A_ = [] for i, row in enumerate(UpperCamelCase__ ): A_ = [] for j, tile in enumerate(UpperCamelCase__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: A_ = self.blend_v(rows[i - 1][j] , UpperCamelCase__ , UpperCamelCase__ ) if j > 0: A_ = self.blend_h(row[j - 1] , UpperCamelCase__ , UpperCamelCase__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(UpperCamelCase__ , dim=3 ) ) A_ = torch.cat(UpperCamelCase__ , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = False , UpperCamelCase__ = True , UpperCamelCase__ = None , ) -> Union[DecoderOutput, torch.FloatTensor]: '''simple docstring''' A_ = sample A_ = self.encode(UpperCamelCase__ ).latent_dist if sample_posterior: A_ = posterior.sample(generator=UpperCamelCase__ ) else: A_ = posterior.mode() A_ = self.decode(UpperCamelCase__ ).sample if not return_dict: return (dec,) return DecoderOutput(sample=UpperCamelCase__ )
162
'''simple docstring''' from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class A__ : lowercase = 42 lowercase = None lowercase = None def UpperCAmelCase__ ( ) -> Node | None: A_ = Node(1 ) A_ = Node(2 ) A_ = Node(3 ) A_ = Node(4 ) A_ = Node(5 ) return tree def UpperCAmelCase__ ( UpperCAmelCase__ ) -> list[int]: return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def UpperCAmelCase__ ( UpperCAmelCase__ ) -> list[int]: return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def UpperCAmelCase__ ( UpperCAmelCase__ ) -> list[int]: return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def UpperCAmelCase__ ( UpperCAmelCase__ ) -> int: return (max(height(root.left ), height(root.right ) ) + 1) if root else 0 def UpperCAmelCase__ ( UpperCAmelCase__ ) -> Sequence[Node | None]: A_ = [] if root is None: return output A_ = deque([root] ) while process_queue: A_ = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__ ) -> Sequence[Node | None]: A_ = [] def populate_output(UpperCAmelCase__, UpperCAmelCase__ ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left, level - 1 ) populate_output(root.right, level - 1 ) populate_output(UpperCAmelCase__, UpperCAmelCase__ ) return output def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__ ) -> Sequence[Node | None]: A_ = [] def populate_output(UpperCAmelCase__, UpperCAmelCase__ ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right, level - 1 ) populate_output(root.left, level - 1 ) populate_output(UpperCAmelCase__, UpperCAmelCase__ ) return output def UpperCAmelCase__ ( UpperCAmelCase__ ) -> Sequence[Node | None] | list[Any]: if root is None: return [] A_ = [] A_ = 0 A_ = height(UpperCAmelCase__ ) for h in range(1, height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(UpperCAmelCase__, UpperCAmelCase__ ) ) A_ = 1 else: output.append(get_nodes_from_right_to_left(UpperCAmelCase__, UpperCAmelCase__ ) ) A_ = 0 return output def UpperCAmelCase__ ( ) -> None: # Main function for testing. A_ = make_tree() print(F'''In-order Traversal: {inorder(UpperCAmelCase__ )}''' ) print(F'''Pre-order Traversal: {preorder(UpperCAmelCase__ )}''' ) print(F'''Post-order Traversal: {postorder(UpperCAmelCase__ )}''', """\n""" ) print(F'''Height of Tree: {height(UpperCAmelCase__ )}''', """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(UpperCAmelCase__ ), """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1, height(UpperCAmelCase__ ) + 1 ): print(F'''Level {level}:''', get_nodes_from_left_to_right(UpperCAmelCase__, level=UpperCAmelCase__ ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(UpperCAmelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
162
1
'''simple docstring''' import os # Precomputes a list of the 100 first triangular numbers __UpperCAmelCase =[int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def __lowerCAmelCase ( ) -> Optional[Any]: __lowerCamelCase = os.path.dirname(os.path.realpath(UpperCamelCase__ ) ) __lowerCamelCase = os.path.join(UpperCamelCase__ , '''words.txt''' ) __lowerCamelCase = '''''' with open(UpperCamelCase__ ) as f: __lowerCamelCase = f.readline() __lowerCamelCase = [word.strip('''"''' ) for word in words.strip('''\r\n''' ).split(''',''' )] __lowerCamelCase = [ word for word in [sum(ord(UpperCamelCase__ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(UpperCamelCase__ ) if __name__ == "__main__": print(solution())
237
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase ={ "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys __UpperCAmelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
237
1