code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def UpperCamelCase ( lowercase_ : Tuple ) -> Optional[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Dict ) -> List[str]: '''simple docstring''' return (-y * np.log(lowercase_ ) - (1 - y) * np.log(1 - h )).mean() def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : Tuple ) -> Union[str, Any]: '''simple docstring''' lowercase =np.dot(lowercase_ , lowercase_ ) return np.sum(y * scores - np.log(1 + np.exp(lowercase_ ) ) ) def UpperCamelCase ( lowercase_ : int , lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[int]=7_0_0_0_0 ) -> str: '''simple docstring''' lowercase =np.zeros(x.shape[1] ) for iterations in range(lowercase_ ): lowercase =np.dot(lowercase_ , lowercase_ ) lowercase =sigmoid_function(lowercase_ ) lowercase =np.dot(x.T , h - y ) / y.size lowercase =theta - alpha * gradient # updating the weights lowercase =np.dot(lowercase_ , lowercase_ ) lowercase =sigmoid_function(lowercase_ ) lowercase =cost_function(lowercase_ , lowercase_ ) if iterations % 1_0_0 == 0: print(f'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": _UpperCAmelCase : List[str] = datasets.load_iris() _UpperCAmelCase : Optional[Any] = iris.data[:, :2] _UpperCAmelCase : Union[str, Any] = (iris.target != 0) * 1 _UpperCAmelCase : List[str] = 0.1 _UpperCAmelCase : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_00_00) print('''theta: ''', theta) # printing the theta i.e our weights vector def UpperCamelCase ( lowercase_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' return sigmoid_function( np.dot(lowercase_ , lowercase_ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((_UpperCAmelCase) , (_UpperCAmelCase)) : List[Any] = (x[:, 0].min(), x[:, 0].max()) ((_UpperCAmelCase) , (_UpperCAmelCase)) : Union[str, Any] = (x[:, 1].min(), x[:, 1].max()) ((_UpperCAmelCase) , (_UpperCAmelCase)) : int = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) _UpperCAmelCase : int = np.c_[xxa.ravel(), xxa.ravel()] _UpperCAmelCase : Union[str, Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
72
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
1
'''simple docstring''' import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class __magic_name__ ( unittest.TestCase ): def _A( self ): lowercase =inspect.getfile(accelerate.test_utils ) lowercase =os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py'''] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 lowercase =test_metrics @require_cpu def _A( self ): debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def _A( self ): debug_launcher(self.test_metrics.main ) @require_single_gpu def _A( self ): self.test_metrics.main() @require_multi_gpu def _A( self ): print(f'Found {torch.cuda.device_count()} devices.' ) lowercase =['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(snake_case_ , env=os.environ.copy() )
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
1
'''simple docstring''' def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return sum(e for e in range(3 , lowercase_ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"""{solution() = }""")
72
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
1
'''simple docstring''' from collections.abc import Sequence from queue import Queue class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None , snake_case_=None ): lowercase =start lowercase =end lowercase =val lowercase =(start + end) // 2 lowercase =left lowercase =right def __repr__( self ): return f'SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})' class __magic_name__ : def __init__( self , snake_case_ , snake_case_ ): lowercase =collection lowercase =function if self.collection: lowercase =self._build_tree(0 , len(snake_case_ ) - 1 ) def _A( self , snake_case_ , snake_case_ ): self._update_tree(self.root , snake_case_ , snake_case_ ) def _A( self , snake_case_ , snake_case_ ): return self._query_range(self.root , snake_case_ , snake_case_ ) def _A( self , snake_case_ , snake_case_ ): if start == end: return SegmentTreeNode(snake_case_ , snake_case_ , self.collection[start] ) lowercase =(start + end) // 2 lowercase =self._build_tree(snake_case_ , snake_case_ ) lowercase =self._build_tree(mid + 1 , snake_case_ ) return SegmentTreeNode(snake_case_ , snake_case_ , self.fn(left.val , right.val ) , snake_case_ , snake_case_ ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): if node.start == i and node.end == i: lowercase =val return if i <= node.mid: self._update_tree(node.left , snake_case_ , snake_case_ ) else: self._update_tree(node.right , snake_case_ , snake_case_ ) lowercase =self.fn(node.left.val , node.right.val ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , snake_case_ , snake_case_ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , snake_case_ , node.mid ) , self._query_range(node.right , node.mid + 1 , snake_case_ ) , ) else: # range in right child tree return self._query_range(node.right , snake_case_ , snake_case_ ) def _A( self ): if self.root is not None: lowercase =Queue() queue.put(self.root ) while not queue.empty(): lowercase =queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 50) _UpperCAmelCase : Any = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
72
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
1
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _UpperCAmelCase : int = '''\ Text data. Second line of data.''' _UpperCAmelCase : Optional[int] = '''file''' @pytest.fixture(scope='''session''' ) def UpperCamelCase ( lowercase_ : str ) -> int: '''simple docstring''' lowercase =tmp_path_factory.mktemp('''data''' ) / (FILE_PATH + '''.zstd''') lowercase =bytes(lowercase_ , '''utf-8''' ) with zstd.open(lowercase_ , '''wb''' ) as f: f.write(lowercase_ ) return path @pytest.fixture def UpperCamelCase ( lowercase_ : Optional[int] ) -> List[Any]: '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir , lowercase_ ) , '''w''' ) as f: f.write(lowercase_ ) return FILE_PATH @pytest.mark.parametrize('''compression_format''' , ['''gzip''', '''xz''', '''zstd'''] ) def UpperCamelCase ( lowercase_ : str , lowercase_ : Dict , lowercase_ : Optional[Any] , lowercase_ : Any , lowercase_ : Optional[Any] , lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase ={'''gzip''': gz_file, '''xz''': xz_file, '''zstd''': zstd_path} lowercase =input_paths[compression_format] lowercase =tmp_path / '''cache''' lowercase =DownloadConfig(cache_dir=lowercase_ , extract_compressed_file=lowercase_ ) lowercase =cached_path(lowercase_ , download_config=lowercase_ ) with open(lowercase_ ) as f: lowercase =f.read() with open(lowercase_ ) as f: lowercase =f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('''default_extracted''' , [True, False] ) @pytest.mark.parametrize('''default_cache_dir''' , [True, False] ) def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : List[Any] ) -> Union[str, Any]: '''simple docstring''' lowercase ='''custom_cache''' lowercase ='''custom_extracted_dir''' lowercase =tmp_path / '''custom_extracted_path''' if default_extracted: lowercase =('''downloads''' if default_cache_dir else custom_cache_dir, '''extracted''') else: monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_DIR''' , lowercase_ ) monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(lowercase_ ) ) lowercase =custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowercase =xz_file lowercase =( DownloadConfig(extract_compressed_file=lowercase_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=lowercase_ ) ) lowercase =cached_path(lowercase_ , download_config=lowercase_ ) assert Path(lowercase_ ).parent.parts[-2:] == expected def UpperCamelCase ( lowercase_ : List[str] ) -> Any: '''simple docstring''' lowercase =str(Path(lowercase_ ).resolve() ) assert cached_path(lowercase_ ) == text_file # relative path lowercase =str(Path(lowercase_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowercase_ ) == text_file def UpperCamelCase ( lowercase_ : int ) -> Dict: '''simple docstring''' lowercase =str(tmp_path.resolve() / '''__missing_file__.txt''' ) with pytest.raises(lowercase_ ): cached_path(lowercase_ ) # relative path lowercase ='''./__missing_file__.txt''' with pytest.raises(lowercase_ ): cached_path(lowercase_ ) def UpperCamelCase ( lowercase_ : List[str] ) -> Tuple: '''simple docstring''' lowercase =get_from_cache(f'tmp://{tmpfs_file}' ) with open(lowercase_ ) as f: lowercase =f.read() assert output_file_content == FILE_CONTENT @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowercase_ ) def UpperCamelCase ( ) -> Union[str, Any]: '''simple docstring''' with pytest.raises(lowercase_ ): cached_path('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowercase_ ) def UpperCamelCase ( lowercase_ : Dict ) -> int: '''simple docstring''' lowercase =tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowercase_ ): http_get('''https://huggingface.co''' , temp_file=lowercase_ ) with pytest.raises(lowercase_ ): http_head('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowercase_ ) def UpperCamelCase ( lowercase_ : Tuple ) -> Union[str, Any]: '''simple docstring''' lowercase =tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowercase_ ): ftp_get('''ftp://huggingface.co''' , temp_file=lowercase_ ) with pytest.raises(lowercase_ ): ftp_head('''ftp://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowercase_ ) def UpperCamelCase ( lowercase_ : Dict ) -> Any: '''simple docstring''' lowercase =tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowercase_ ): fsspec_get('''s3://huggingface.co''' , temp_file=lowercase_ ) with pytest.raises(lowercase_ ): fsspec_head('''s3://huggingface.co''' )
72
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' from __future__ import annotations import math def UpperCamelCase ( lowercase_ : int , lowercase_ : int , lowercase_ : bool , lowercase_ : list[int] , lowercase_ : float ) -> int: '''simple docstring''' if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if len(lowercase_ ) == 0: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , lowercase_ , lowercase_ , lowercase_ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase_ , lowercase_ , lowercase_ ) , ) return min( minimax(depth + 1 , node_index * 2 , lowercase_ , lowercase_ , lowercase_ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase_ , lowercase_ , lowercase_ ) , ) def UpperCamelCase ( ) -> None: '''simple docstring''' lowercase =[9_0, 2_3, 6, 3_3, 2_1, 6_5, 1_2_3, 3_4_4_2_3] lowercase =math.log(len(lowercase_ ) , 2 ) print('''Optimal value : ''' , end='''''' ) print(minimax(0 , 0 , lowercase_ , lowercase_ , lowercase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
72
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
1
'''simple docstring''' from ...processing_utils import ProcessorMixin class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'SpeechT5FeatureExtractor' UpperCamelCase__ = 'SpeechT5Tokenizer' def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): lowercase =kwargs.pop('''audio''' , snake_case_ ) lowercase =kwargs.pop('''text''' , snake_case_ ) lowercase =kwargs.pop('''text_target''' , snake_case_ ) lowercase =kwargs.pop('''audio_target''' , snake_case_ ) lowercase =kwargs.pop('''sampling_rate''' , snake_case_ ) if audio is not None and text is not None: raise ValueError( '''Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?''' ) if audio_target is not None and text_target is not None: raise ValueError( '''Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?''' ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( '''You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.''' ) if audio is not None: lowercase =self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) elif text is not None: lowercase =self.tokenizer(snake_case_ , **snake_case_ ) else: lowercase =None if audio_target is not None: lowercase =self.feature_extractor(audio_target=snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) lowercase =targets['''input_values'''] elif text_target is not None: lowercase =self.tokenizer(snake_case_ , **snake_case_ ) lowercase =targets['''input_ids'''] else: lowercase =None if inputs is None: return targets if targets is not None: lowercase =labels lowercase =targets.get('''attention_mask''' ) if decoder_attention_mask is not None: lowercase =decoder_attention_mask return inputs def _A( self , *snake_case_ , **snake_case_ ): lowercase =kwargs.pop('''input_values''' , snake_case_ ) lowercase =kwargs.pop('''input_ids''' , snake_case_ ) lowercase =kwargs.pop('''labels''' , snake_case_ ) if input_values is not None and input_ids is not None: raise ValueError('''Cannot process both `input_values` and `input_ids` inputs.''' ) if input_values is None and input_ids is None and labels is None: raise ValueError( '''You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.''' ) if input_values is not None: lowercase =self.feature_extractor.pad(snake_case_ , *snake_case_ , **snake_case_ ) elif input_ids is not None: lowercase =self.tokenizer.pad(snake_case_ , **snake_case_ ) else: lowercase =None if labels is not None: if "input_ids" in labels or (isinstance(snake_case_ , snake_case_ ) and "input_ids" in labels[0]): lowercase =self.tokenizer.pad(snake_case_ , **snake_case_ ) lowercase =targets['''input_ids'''] else: lowercase =self.feature_extractor.feature_size lowercase =self.feature_extractor.num_mel_bins lowercase =self.feature_extractor.pad(snake_case_ , *snake_case_ , **snake_case_ ) lowercase =feature_size_hack lowercase =targets['''input_values'''] else: lowercase =None if inputs is None: return targets if targets is not None: lowercase =labels lowercase =targets.get('''attention_mask''' ) if decoder_attention_mask is not None: lowercase =decoder_attention_mask return inputs def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ )
72
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
1
'''simple docstring''' import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _UpperCAmelCase : Any = '''src/diffusers''' _UpperCAmelCase : str = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _UpperCAmelCase : List[str] = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _UpperCAmelCase : Optional[Any] = spec.loader.load_module() def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : int ) -> List[Any]: '''simple docstring''' return line.startswith(lowercase_ ) or len(lowercase_ ) <= 1 or re.search(R'''^\s*\)(\s*->.*:|:)\s*$''' , lowercase_ ) is not None def UpperCamelCase ( lowercase_ : List[Any] ) -> Any: '''simple docstring''' lowercase =object_name.split('''.''' ) lowercase =0 # First let's find the module where our object lives. lowercase =parts[i] while i < len(lowercase_ ) and not os.path.isfile(os.path.join(lowercase_ , f'{module}.py' ) ): i += 1 if i < len(lowercase_ ): lowercase =os.path.join(lowercase_ , parts[i] ) if i >= len(lowercase_ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(lowercase_ , f'{module}.py' ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase =f.readlines() # Now let's find the class / func in the code! lowercase ='''''' lowercase =0 for name in parts[i + 1 :]: while ( line_index < len(lowercase_ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowercase_ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowercase =line_index while line_index < len(lowercase_ ) and _should_continue(lines[line_index] , lowercase_ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase =lines[start_index:line_index] return "".join(lowercase_ ) _UpperCAmelCase : Any = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _UpperCAmelCase : Optional[int] = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') _UpperCAmelCase : Optional[int] = re.compile(r'''<FILL\s+[^>]*>''') def UpperCamelCase ( lowercase_ : List[str] ) -> Optional[int]: '''simple docstring''' lowercase =code.split('''\n''' ) lowercase =0 while idx < len(lowercase_ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowercase_ ): return re.search(R'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def UpperCamelCase ( lowercase_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' lowercase =len(get_indent(lowercase_ ) ) > 0 if has_indent: lowercase =f'class Bla:\n{code}' lowercase =black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_1_9 , preview=lowercase_ ) lowercase =black.format_str(lowercase_ , mode=lowercase_ ) lowercase , lowercase =style_docstrings_in_code(lowercase_ ) return result[len('''class Bla:\n''' ) :] if has_indent else result def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : int=False ) -> int: '''simple docstring''' with open(lowercase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase =f.readlines() lowercase =[] lowercase =0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowercase_ ): lowercase =_re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowercase , lowercase , lowercase =search.groups() lowercase =find_code_in_diffusers(lowercase_ ) lowercase =get_indent(lowercase_ ) lowercase =line_index + 1 if indent == theoretical_indent else line_index + 2 lowercase =theoretical_indent lowercase =start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowercase =True while line_index < len(lowercase_ ) and should_continue: line_index += 1 if line_index >= len(lowercase_ ): break lowercase =lines[line_index] lowercase =_should_continue(lowercase_ , lowercase_ ) and re.search(f'^{indent}# End copy' , lowercase_ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase =lines[start_index:line_index] lowercase =''''''.join(lowercase_ ) # Remove any nested `Copied from` comments to avoid circular copies lowercase =[line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(lowercase_ ) is None] lowercase ='''\n'''.join(lowercase_ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowercase_ ) > 0: lowercase =replace_pattern.replace('''with''' , '''''' ).split(''',''' ) lowercase =[_re_replace_pattern.search(lowercase_ ) for p in patterns] for pattern in patterns: if pattern is None: continue lowercase , lowercase , lowercase =pattern.groups() lowercase =re.sub(lowercase_ , lowercase_ , lowercase_ ) if option.strip() == "all-casing": lowercase =re.sub(obja.lower() , obja.lower() , lowercase_ ) lowercase =re.sub(obja.upper() , obja.upper() , lowercase_ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowercase =blackify(lines[start_index - 1] + theoretical_code ) lowercase =theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowercase =lines[:start_index] + [theoretical_code] + lines[line_index:] lowercase =start_index + 1 if overwrite and len(lowercase_ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(lowercase_ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lowercase_ ) return diffs def UpperCamelCase ( lowercase_ : bool = False ) -> Dict: '''simple docstring''' lowercase =glob.glob(os.path.join(lowercase_ , '''**/*.py''' ) , recursive=lowercase_ ) lowercase =[] for filename in all_files: lowercase =is_copy_consistent(lowercase_ , lowercase_ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(lowercase_ ) > 0: lowercase ='''\n'''.join(lowercase_ ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _UpperCAmelCase : Dict = parser.parse_args() check_copies(args.fix_and_overwrite)
72
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
1
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
1
'''simple docstring''' _UpperCAmelCase : int = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _UpperCAmelCase : Dict = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] _UpperCAmelCase : Optional[int] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin _UpperCAmelCase : Dict = logging.get_logger(__name__) enable_full_determinism() class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetaDModel UpperCamelCase__ = 'sample' @property def _A( self ): lowercase =4 lowercase =3 lowercase =(32, 32) lowercase =floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) lowercase =torch.tensor([10] ).to(snake_case_ ) return {"sample": noise, "timestep": time_step} @property def _A( self ): return (3, 32, 32) @property def _A( self ): return (3, 32, 32) def _A( self ): lowercase ={ '''block_out_channels''': (32, 64), '''down_block_types''': ('''DownBlock2D''', '''AttnDownBlock2D'''), '''up_block_types''': ('''AttnUpBlock2D''', '''UpBlock2D'''), '''attention_head_dim''': 3, '''out_channels''': 3, '''in_channels''': 3, '''layers_per_block''': 2, '''sample_size''': 32, } lowercase =self.dummy_input return init_dict, inputs_dict class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetaDModel UpperCamelCase__ = 'sample' @property def _A( self ): lowercase =4 lowercase =4 lowercase =(32, 32) lowercase =floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) lowercase =torch.tensor([10] ).to(snake_case_ ) return {"sample": noise, "timestep": time_step} @property def _A( self ): return (4, 32, 32) @property def _A( self ): return (4, 32, 32) def _A( self ): lowercase ={ '''sample_size''': 32, '''in_channels''': 4, '''out_channels''': 4, '''layers_per_block''': 2, '''block_out_channels''': (32, 64), '''attention_head_dim''': 32, '''down_block_types''': ('''DownBlock2D''', '''DownBlock2D'''), '''up_block_types''': ('''UpBlock2D''', '''UpBlock2D'''), } lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase , lowercase =UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=snake_case_ ) self.assertIsNotNone(snake_case_ ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(snake_case_ ) lowercase =model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != '''cuda''' , '''This test is supposed to run on GPU''' ) def _A( self ): lowercase , lowercase =UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=snake_case_ ) model.to(snake_case_ ) lowercase =model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != '''cuda''' , '''This test is supposed to run on GPU''' ) def _A( self ): # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` lowercase , lowercase =UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=snake_case_ ) model_accelerate.to(snake_case_ ) model_accelerate.eval() lowercase =torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) lowercase =noise.to(snake_case_ ) lowercase =torch.tensor([10] * noise.shape[0] ).to(snake_case_ ) lowercase =model_accelerate(snake_case_ , snake_case_ )['''sample'''] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() lowercase , lowercase =UNetaDModel.from_pretrained( '''fusing/unet-ldm-dummy-update''' , output_loading_info=snake_case_ , low_cpu_mem_usage=snake_case_ ) model_normal_load.to(snake_case_ ) model_normal_load.eval() lowercase =model_normal_load(snake_case_ , snake_case_ )['''sample'''] assert torch_all_close(snake_case_ , snake_case_ , rtol=1E-3 ) def _A( self ): lowercase =UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' ) model.eval() model.to(snake_case_ ) lowercase =torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) lowercase =noise.to(snake_case_ ) lowercase =torch.tensor([10] * noise.shape[0] ).to(snake_case_ ) with torch.no_grad(): lowercase =model(snake_case_ , snake_case_ ).sample lowercase =output[0, -1, -3:, -3:].flatten().cpu() # fmt: off lowercase =torch.tensor([-13.32_58, -20.11_00, -15.98_73, -17.66_17, -23.05_96, -17.94_19, -13.36_75, -16.18_89, -12.38_00] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1E-3 ) ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetaDModel UpperCamelCase__ = 'sample' @property def _A( self , snake_case_=(32, 32) ): lowercase =4 lowercase =3 lowercase =floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) lowercase =torch.tensor(batch_size * [10] ).to(dtype=torch.intaa , device=snake_case_ ) return {"sample": noise, "timestep": time_step} @property def _A( self ): return (3, 32, 32) @property def _A( self ): return (3, 32, 32) def _A( self ): lowercase ={ '''block_out_channels''': [32, 64, 64, 64], '''in_channels''': 3, '''layers_per_block''': 1, '''out_channels''': 3, '''time_embedding_type''': '''fourier''', '''norm_eps''': 1E-6, '''mid_block_scale_factor''': math.sqrt(2.0 ), '''norm_num_groups''': None, '''down_block_types''': [ '''SkipDownBlock2D''', '''AttnSkipDownBlock2D''', '''SkipDownBlock2D''', '''SkipDownBlock2D''', ], '''up_block_types''': [ '''SkipUpBlock2D''', '''SkipUpBlock2D''', '''AttnSkipUpBlock2D''', '''SkipUpBlock2D''', ], } lowercase =self.dummy_input return init_dict, inputs_dict @slow def _A( self ): lowercase , lowercase =UNetaDModel.from_pretrained('''google/ncsnpp-celebahq-256''' , output_loading_info=snake_case_ ) self.assertIsNotNone(snake_case_ ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(snake_case_ ) lowercase =self.dummy_input lowercase =floats_tensor((4, 3) + (2_56, 2_56) ).to(snake_case_ ) lowercase =noise lowercase =model(**snake_case_ ) assert image is not None, "Make sure output is not None" @slow def _A( self ): lowercase =UNetaDModel.from_pretrained('''google/ncsnpp-celebahq-256''' ) model.to(snake_case_ ) lowercase =4 lowercase =3 lowercase =(2_56, 2_56) lowercase =torch.ones((batch_size, num_channels) + sizes ).to(snake_case_ ) lowercase =torch.tensor(batch_size * [1E-4] ).to(snake_case_ ) with torch.no_grad(): lowercase =model(snake_case_ , snake_case_ ).sample lowercase =output[0, -3:, -3:, -1].flatten().cpu() # fmt: off lowercase =torch.tensor([-48_42.86_91, -64_99.66_31, -38_00.19_53, -79_78.26_86, -1_09_80.71_29, -2_00_28.85_35, 81_48.28_22, 23_42.29_05, 5_67.76_08] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1E-2 ) ) def _A( self ): lowercase =UNetaDModel.from_pretrained('''fusing/ncsnpp-ffhq-ve-dummy-update''' ) model.to(snake_case_ ) lowercase =4 lowercase =3 lowercase =(32, 32) lowercase =torch.ones((batch_size, num_channels) + sizes ).to(snake_case_ ) lowercase =torch.tensor(batch_size * [1E-4] ).to(snake_case_ ) with torch.no_grad(): lowercase =model(snake_case_ , snake_case_ ).sample lowercase =output[0, -3:, -3:, -1].flatten().cpu() # fmt: off lowercase =torch.tensor([-0.03_25, -0.09_00, -0.08_69, -0.03_32, -0.07_25, -0.02_70, -0.01_01, 0.02_27, 0.02_56] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1E-2 ) ) def _A( self ): # not required for this model pass
72
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = DanceDiffusionPipeline UpperCamelCase__ = UNCONDITIONAL_AUDIO_GENERATION_PARAMS UpperCamelCase__ = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } UpperCamelCase__ = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): torch.manual_seed(0 ) lowercase =UNetaDModel( block_out_channels=(32, 32, 64) , extra_in_channels=16 , sample_size=5_12 , sample_rate=1_60_00 , in_channels=2 , out_channels=2 , flip_sin_to_cos=snake_case_ , use_timestep_embedding=snake_case_ , time_embedding_type='''fourier''' , mid_block_type='''UNetMidBlock1D''' , down_block_types=('''DownBlock1DNoSkip''', '''DownBlock1D''', '''AttnDownBlock1D''') , up_block_types=('''AttnUpBlock1D''', '''UpBlock1D''', '''UpBlock1DNoSkip''') , ) lowercase =IPNDMScheduler() lowercase ={ '''unet''': unet, '''scheduler''': scheduler, } return components def _A( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith('''mps''' ): lowercase =torch.manual_seed(snake_case_ ) else: lowercase =torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) lowercase ={ '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 4, } return inputs def _A( self ): lowercase ='''cpu''' # ensure determinism for the device-dependent torch.Generator lowercase =self.get_dummy_components() lowercase =DanceDiffusionPipeline(**snake_case_ ) lowercase =pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) lowercase =self.get_dummy_inputs(snake_case_ ) lowercase =pipe(**snake_case_ ) lowercase =output.audios lowercase =audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) lowercase =np.array([-0.72_65, 1.00_00, -0.83_88, 0.11_75, 0.94_98, -1.00_00] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def _A( self ): return super().test_save_load_local() @skip_mps def _A( self ): return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) @skip_mps def _A( self ): return super().test_save_load_optional_components() @skip_mps def _A( self ): return super().test_attention_slicing_forward_pass() def _A( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase ): def _A( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _A( self ): lowercase =torch_device lowercase =DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' ) lowercase =pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) lowercase =torch.manual_seed(0 ) lowercase =pipe(generator=snake_case_ , num_inference_steps=1_00 , audio_length_in_s=4.0_96 ) lowercase =output.audios lowercase =audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) lowercase =np.array([-0.01_92, -0.02_31, -0.03_18, -0.00_59, 0.00_02, -0.00_20] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 def _A( self ): lowercase =torch_device lowercase =DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' , torch_dtype=torch.floataa ) lowercase =pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) lowercase =torch.manual_seed(0 ) lowercase =pipe(generator=snake_case_ , num_inference_steps=1_00 , audio_length_in_s=4.0_96 ) lowercase =output.audios lowercase =audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) lowercase =np.array([-0.03_67, -0.04_88, -0.07_71, -0.05_25, -0.04_44, -0.03_41] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2
72
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
1
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'ChineseCLIPImageProcessor' UpperCamelCase__ = ('BertTokenizer', 'BertTokenizerFast') def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): lowercase =None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , snake_case_ , ) lowercase =kwargs.pop('''feature_extractor''' ) lowercase =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(snake_case_ , snake_case_ ) lowercase =self.image_processor def __call__( self , snake_case_=None , snake_case_=None , snake_case_=None , **snake_case_ ): if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: lowercase =self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if text is not None and images is not None: lowercase =image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case_ ) , tensor_type=snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _A( self ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , snake_case_ , ) return self.image_processor_class
72
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
1
'''simple docstring''' # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar _UpperCAmelCase : int = TypeVar('''T''') class __magic_name__ ( Generic[T] ): def __init__( self , snake_case_ = True ): lowercase ={} # dictionary of lists lowercase =directed def _A( self , snake_case_ , snake_case_ ): if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case_ ) self.adj_list[destination_vertex].append(snake_case_ ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case_ ) lowercase =[source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(snake_case_ ) lowercase =[destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: lowercase =[destination_vertex] lowercase =[source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case_ ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case_ ) lowercase =[] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: lowercase =[destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: lowercase =[destination_vertex] lowercase =[] return self def __repr__( self ): return pformat(self.adj_list )
72
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : str = { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/config.json''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/config.json''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'xlm-roberta' def __init__( self , snake_case_=3_05_22 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_=30_72 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , **snake_case_ , ): super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) lowercase =vocab_size lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =hidden_act lowercase =intermediate_size lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =max_position_embeddings lowercase =type_vocab_size lowercase =initializer_range lowercase =layer_norm_eps lowercase =position_embedding_type lowercase =use_cache lowercase =classifier_dropout class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property def _A( self ): if self.task == "multiple-choice": lowercase ={0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase ={0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
72
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
1
'''simple docstring''' from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class __magic_name__ : UpperCamelCase__ = 42 UpperCamelCase__ = 42 class __magic_name__ : def __init__( self , snake_case_ ): lowercase =[[] for _ in range(snake_case_ )] lowercase =size def __getitem__( self , snake_case_ ): return iter(self._graph[vertex] ) @property def _A( self ): return self._size def _A( self , snake_case_ , snake_case_ , snake_case_ ): if weight not in (0, 1): raise ValueError('''Edge weight must be either 0 or 1.''' ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError('''Vertex indexes must be in [0; size).''' ) self._graph[from_vertex].append(Edge(snake_case_ , snake_case_ ) ) def _A( self , snake_case_ , snake_case_ ): lowercase =deque([start_vertex] ) lowercase =[None] * self.size lowercase =0 while queue: lowercase =queue.popleft() lowercase =distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: lowercase =current_distance + edge.weight lowercase =distances[edge.destination_vertex] if ( isinstance(snake_case_ , snake_case_ ) and new_distance >= dest_vertex_distance ): continue lowercase =new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError('''No path from start_vertex to finish_vertex.''' ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
72
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
1
'''simple docstring''' import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : List[Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : List[Any] = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } _UpperCAmelCase : Any = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) _UpperCAmelCase : Any = 0 _UpperCAmelCase : List[Any] = 1 _UpperCAmelCase : str = 2 _UpperCAmelCase : Dict = 3 _UpperCAmelCase : Optional[int] = 4 class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = 'left' def __init__( self , snake_case_ , snake_case_=False , snake_case_=True , snake_case_=False , snake_case_="<s>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<sep>" , snake_case_="<pad>" , snake_case_="<cls>" , snake_case_="<mask>" , snake_case_=["<eop>", "<eod>"] , snake_case_ = None , **snake_case_ , ): # Mask token behave like a normal word, i.e. include the space before it lowercase =AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case_ , remove_space=snake_case_ , keep_accents=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , pad_token=snake_case_ , cls_token=snake_case_ , mask_token=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =3 lowercase =do_lower_case lowercase =remove_space lowercase =keep_accents lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) @property def _A( self ): return len(self.sp_model ) def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): if self.remove_space: lowercase =''' '''.join(inputs.strip().split() ) else: lowercase =inputs lowercase =outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: lowercase =unicodedata.normalize('''NFKD''' , snake_case_ ) lowercase =''''''.join([c for c in outputs if not unicodedata.combining(snake_case_ )] ) if self.do_lower_case: lowercase =outputs.lower() return outputs def _A( self , snake_case_ ): lowercase =self.preprocess_text(snake_case_ ) lowercase =self.sp_model.encode(snake_case_ , out_type=snake_case_ ) lowercase =[] for piece in pieces: if len(snake_case_ ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): lowercase =self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case_ , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowercase =cur_pieces[1:] else: lowercase =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case_ ) else: new_pieces.append(snake_case_ ) return new_pieces def _A( self , snake_case_ ): return self.sp_model.PieceToId(snake_case_ ) def _A( self , snake_case_ ): return self.sp_model.IdToPiece(snake_case_ ) def _A( self , snake_case_ ): lowercase =''''''.join(snake_case_ ).replace(snake_case_ , ''' ''' ).strip() return out_string def _A( self , snake_case_ , snake_case_ = False , snake_case_ = None , snake_case_ = True , **snake_case_ , ): lowercase =kwargs.pop('''use_source_tokenizer''' , snake_case_ ) lowercase =self.convert_ids_to_tokens(snake_case_ , skip_special_tokens=snake_case_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowercase =[] lowercase =[] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case_ ) ) lowercase =[] sub_texts.append(snake_case_ ) else: current_sub_text.append(snake_case_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case_ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowercase =''''''.join(snake_case_ ) lowercase =( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowercase =self.clean_up_tokenization(snake_case_ ) return clean_text else: return text def _A( self , snake_case_ , snake_case_ = None ): lowercase =[self.sep_token_id] lowercase =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) if token_ids_a is not None: return ([0] * len(snake_case_ )) + [1] + ([0] * len(snake_case_ )) + [1, 1] return ([0] * len(snake_case_ )) + [1, 1] def _A( self , snake_case_ , snake_case_ = None ): lowercase =[self.sep_token_id] lowercase =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
1
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { '''roberta-base''': '''https://huggingface.co/roberta-base/resolve/main/config.json''', '''roberta-large''': '''https://huggingface.co/roberta-large/resolve/main/config.json''', '''roberta-large-mnli''': '''https://huggingface.co/roberta-large-mnli/resolve/main/config.json''', '''distilroberta-base''': '''https://huggingface.co/distilroberta-base/resolve/main/config.json''', '''roberta-base-openai-detector''': '''https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json''', '''roberta-large-openai-detector''': '''https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'roberta' def __init__( self , snake_case_=5_02_65 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_=30_72 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , **snake_case_ , ): super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) lowercase =vocab_size lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =hidden_act lowercase =intermediate_size lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =max_position_embeddings lowercase =type_vocab_size lowercase =initializer_range lowercase =layer_norm_eps lowercase =position_embedding_type lowercase =use_cache lowercase =classifier_dropout class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property def _A( self ): if self.task == "multiple-choice": lowercase ={0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase ={0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
72
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
1
'''simple docstring''' def UpperCamelCase ( lowercase_ : int = 5_0 ) -> int: '''simple docstring''' lowercase =[1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Tuple = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys _UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
1
'''simple docstring''' import math def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Any: '''simple docstring''' if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError('''This should never happen''' ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. _UpperCAmelCase : Optional[Any] = '''Enter the base and the power separated by a comma: ''' _UpperCAmelCase , _UpperCAmelCase : Any = map(int, input(prompt).split(''',''')) _UpperCAmelCase , _UpperCAmelCase : str = map(int, input(prompt).split(''',''')) # We find the log of each number, using the function res(), which takes two # arguments. _UpperCAmelCase : Dict = res(xa, ya) _UpperCAmelCase : int = res(xa, ya) # We check for the largest number if resa > resa: print('''Largest number is''', xa, '''^''', ya) elif resa > resa: print('''Largest number is''', xa, '''^''', ya) else: print('''Both are equal''')
72
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--user''', type=str, default='''ubuntu''') parser.add_argument('''--host''', type=str, default='''localhost''') parser.add_argument('''--key_path''', type=str, default=None) parser.add_argument('''--instance''', type=str, default='''V100:1''') parser.add_argument('''--provider''', type=str, default='''cheapest''') parser.add_argument('''--use_spot''', type=bool, default=False) parser.add_argument('''--example''', type=str, default='''pytorch/text-generation/run_generation.py''') _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('''Cannot specify both BYO and on-demand cluster args''') _UpperCAmelCase : Dict = rh.cluster( name='''rh-cluster''', ips=[args.host], ssh_creds={'''ssh_user''': args.user, '''ssh_private_key''': args.key_path} ) else: _UpperCAmelCase : Any = rh.cluster( name='''rh-cluster''', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) _UpperCAmelCase : List[str] = args.example.rsplit('''/''', 1)[0] # Set up remote environment cluster.install_packages(['''pip:./''']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F"""pip install -r transformers/examples/{example_dir}/requirements.txt"""]) cluster.run(['''pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117''']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F"""python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}"""]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
72
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
1
'''simple docstring''' from math import ceil, sqrt def UpperCamelCase ( lowercase_ : int = 1_0_0_0_0_0_0 ) -> int: '''simple docstring''' lowercase =0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: lowercase =max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: lowercase =1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F"""{solution() = }""")
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { '''microsoft/swin-tiny-patch4-window7-224''': ( '''https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json''' ), # See all Swin models at https://huggingface.co/models?filter=swin } class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'swin' UpperCamelCase__ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=2_24 , snake_case_=4 , snake_case_=3 , snake_case_=96 , snake_case_=[2, 2, 6, 2] , snake_case_=[3, 6, 12, 24] , snake_case_=7 , snake_case_=4.0 , snake_case_=True , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.1 , snake_case_="gelu" , snake_case_=False , snake_case_=0.02 , snake_case_=1E-5 , snake_case_=32 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =image_size lowercase =patch_size lowercase =num_channels lowercase =embed_dim lowercase =depths lowercase =len(snake_case_ ) lowercase =num_heads lowercase =window_size lowercase =mlp_ratio lowercase =qkv_bias lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =drop_path_rate lowercase =hidden_act lowercase =use_absolute_embeddings lowercase =layer_norm_eps lowercase =initializer_range lowercase =encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowercase =int(embed_dim * 2 ** (len(snake_case_ ) - 1) ) lowercase =['''stem'''] + [f'stage{idx}' for idx in range(1 , len(snake_case_ ) + 1 )] lowercase , lowercase =get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = version.parse('1.11' ) @property def _A( self ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def _A( self ): return 1E-4
72
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: _UpperCAmelCase : Optional[Any] = None _UpperCAmelCase : Tuple = logging.get_logger(__name__) _UpperCAmelCase : Any = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} _UpperCAmelCase : str = { '''vocab_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json''' ), }, } _UpperCAmelCase : List[str] = { '''moussaKam/mbarthez''': 10_24, '''moussaKam/barthez''': 10_24, '''moussaKam/barthez-orangesum-title''': 10_24, } _UpperCAmelCase : Dict = '''▁''' class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] UpperCamelCase__ = BarthezTokenizer def __init__( self , snake_case_=None , snake_case_=None , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<s>" , snake_case_="<unk>" , snake_case_="<pad>" , snake_case_="<mask>" , **snake_case_ , ): # Mask token behave like a normal word, i.e. include the space before it lowercase =AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token super().__init__( snake_case_ , tokenizer_file=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , cls_token=snake_case_ , pad_token=snake_case_ , mask_token=snake_case_ , **snake_case_ , ) lowercase =vocab_file lowercase =False if not self.vocab_file else True def _A( self , snake_case_ , snake_case_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase =[self.cls_token_id] lowercase =[self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _A( self , snake_case_ , snake_case_ = None ): lowercase =[self.sep_token_id] lowercase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _A( self , snake_case_ , snake_case_ = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ): copyfile(self.vocab_file , snake_case_ ) return (out_vocab_file,)
72
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
1
'''simple docstring''' # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def UpperCamelCase ( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[int] , lowercase_ : Dict ) -> Optional[Any]: '''simple docstring''' lowercase ={ '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, nicht wahr?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase ={ '''wmt16-en-de-dist-12-1''': [2_8.3, 2_7.5_2], '''wmt16-en-de-dist-6-1''': [2_7.4, 2_7.1_1], '''wmt16-en-de-12-1''': [2_6.9, 2_5.7_5], } lowercase =f'{src_lang}-{tgt_lang}' lowercase =f'\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "allenai/{model_name}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don\'t use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n' model_card_dir.mkdir(parents=lowercase_ , exist_ok=lowercase_ ) lowercase =os.path.join(lowercase_ , '''README.md''' ) print(f'Generating {path}' ) with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(lowercase_ ) # make sure we are under the root of the project _UpperCAmelCase : Optional[Any] = Path(__file__).resolve().parent.parent.parent _UpperCAmelCase : Tuple = repo_dir / '''model_cards''' for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: _UpperCAmelCase : str = model_cards_dir / '''allenai''' / model_name write_model_card(model_card_dir, src_lang='''en''', tgt_lang='''de''', model_name=model_name)
72
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' def UpperCamelCase ( lowercase_ : int = 1_0_0 ) -> int: '''simple docstring''' lowercase =0 lowercase =0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
72
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
1
'''simple docstring''' import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def UpperCamelCase ( lowercase_ : int ) -> List[str]: '''simple docstring''' random.seed(lowercase_ ) np.random.seed(lowercase_ ) torch.manual_seed(lowercase_ ) torch.cuda.manual_seed_all(lowercase_ ) # ^^ safe to call this function even if cuda is not available class __magic_name__ : def __init__( self , snake_case_ , snake_case_ = 0.99_99 , snake_case_ = 0.0 , snake_case_ = 0 , snake_case_ = False , snake_case_ = 1.0 , snake_case_ = 2 / 3 , snake_case_ = None , snake_case_ = None , **snake_case_ , ): if isinstance(snake_case_ , torch.nn.Module ): lowercase =( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage`''' , '''1.0.0''' , snake_case_ , standard_warn=snake_case_ , ) lowercase =parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility lowercase =True if kwargs.get('''max_value''' , snake_case_ ) is not None: lowercase ='''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('''max_value''' , '''1.0.0''' , snake_case_ , standard_warn=snake_case_ ) lowercase =kwargs['''max_value'''] if kwargs.get('''min_value''' , snake_case_ ) is not None: lowercase ='''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('''min_value''' , '''1.0.0''' , snake_case_ , standard_warn=snake_case_ ) lowercase =kwargs['''min_value'''] lowercase =list(snake_case_ ) lowercase =[p.clone().detach() for p in parameters] if kwargs.get('''device''' , snake_case_ ) is not None: lowercase ='''The `device` argument is deprecated. Please use `to` instead.''' deprecate('''device''' , '''1.0.0''' , snake_case_ , standard_warn=snake_case_ ) self.to(device=kwargs['''device'''] ) lowercase =None lowercase =decay lowercase =min_decay lowercase =update_after_step lowercase =use_ema_warmup lowercase =inv_gamma lowercase =power lowercase =0 lowercase =None # set in `step()` lowercase =model_cls lowercase =model_config @classmethod def _A( cls , snake_case_ , snake_case_ ): lowercase , lowercase =model_cls.load_config(snake_case_ , return_unused_kwargs=snake_case_ ) lowercase =model_cls.from_pretrained(snake_case_ ) lowercase =cls(model.parameters() , model_cls=snake_case_ , model_config=model.config ) ema_model.load_state_dict(snake_case_ ) return ema_model def _A( self , snake_case_ ): if self.model_cls is None: raise ValueError('''`save_pretrained` can only be used if `model_cls` was defined at __init__.''' ) if self.model_config is None: raise ValueError('''`save_pretrained` can only be used if `model_config` was defined at __init__.''' ) lowercase =self.model_cls.from_config(self.model_config ) lowercase =self.state_dict() state_dict.pop('''shadow_params''' , snake_case_ ) model.register_to_config(**snake_case_ ) self.copy_to(model.parameters() ) model.save_pretrained(snake_case_ ) def _A( self , snake_case_ ): lowercase =max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: lowercase =1 - (1 + step / self.inv_gamma) ** -self.power else: lowercase =(1 + step) / (10 + step) lowercase =min(snake_case_ , self.decay ) # make sure decay is not smaller than min_decay lowercase =max(snake_case_ , self.min_decay ) return cur_decay_value @torch.no_grad() def _A( self , snake_case_ ): if isinstance(snake_case_ , torch.nn.Module ): lowercase =( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage.step`''' , '''1.0.0''' , snake_case_ , standard_warn=snake_case_ , ) lowercase =parameters.parameters() lowercase =list(snake_case_ ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. lowercase =self.get_decay(self.optimization_step ) lowercase =decay lowercase =1 - decay lowercase =contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , snake_case_ ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): lowercase =deepspeed.zero.GatheredParameters(snake_case_ , modifier_rank=snake_case_ ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(snake_case_ ) def _A( self , snake_case_ ): lowercase =list(snake_case_ ) for s_param, param in zip(self.shadow_params , snake_case_ ): param.data.copy_(s_param.to(param.device ).data ) def _A( self , snake_case_=None , snake_case_=None ): lowercase =[ p.to(device=snake_case_ , dtype=snake_case_ ) if p.is_floating_point() else p.to(device=snake_case_ ) for p in self.shadow_params ] def _A( self ): return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def _A( self , snake_case_ ): lowercase =[param.detach().cpu().clone() for param in parameters] def _A( self , snake_case_ ): if self.temp_stored_params is None: raise RuntimeError('''This ExponentialMovingAverage has no `store()`ed weights ''' '''to `restore()`''' ) for c_param, param in zip(self.temp_stored_params , snake_case_ ): param.data.copy_(c_param.data ) # Better memory-wise. lowercase =None def _A( self , snake_case_ ): lowercase =copy.deepcopy(snake_case_ ) lowercase =state_dict.get('''decay''' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('''Decay must be between 0 and 1''' ) lowercase =state_dict.get('''min_decay''' , self.min_decay ) if not isinstance(self.min_decay , snake_case_ ): raise ValueError('''Invalid min_decay''' ) lowercase =state_dict.get('''optimization_step''' , self.optimization_step ) if not isinstance(self.optimization_step , snake_case_ ): raise ValueError('''Invalid optimization_step''' ) lowercase =state_dict.get('''update_after_step''' , self.update_after_step ) if not isinstance(self.update_after_step , snake_case_ ): raise ValueError('''Invalid update_after_step''' ) lowercase =state_dict.get('''use_ema_warmup''' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , snake_case_ ): raise ValueError('''Invalid use_ema_warmup''' ) lowercase =state_dict.get('''inv_gamma''' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('''Invalid inv_gamma''' ) lowercase =state_dict.get('''power''' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('''Invalid power''' ) lowercase =state_dict.get('''shadow_params''' , snake_case_ ) if shadow_params is not None: lowercase =shadow_params if not isinstance(self.shadow_params , snake_case_ ): raise ValueError('''shadow_params must be a list''' ) if not all(isinstance(snake_case_ , torch.Tensor ) for p in self.shadow_params ): raise ValueError('''shadow_params must all be Tensors''' )
72
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor'] UpperCamelCase__ = 'SamImageProcessor' def __init__( self , snake_case_ ): super().__init__(snake_case_ ) lowercase =self.image_processor lowercase =-10 lowercase =self.image_processor.size['''longest_edge'''] def __call__( self , snake_case_=None , snake_case_=None , snake_case_=None , snake_case_=None , snake_case_ = None , **snake_case_ , ): lowercase =self.image_processor( snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # pop arguments that are not used in the foward but used nevertheless lowercase =encoding_image_processor['''original_sizes'''] if hasattr(snake_case_ , '''numpy''' ): # Checks if Torch or TF tensor lowercase =original_sizes.numpy() lowercase , lowercase , lowercase =self._check_and_preprocess_points( input_points=snake_case_ , input_labels=snake_case_ , input_boxes=snake_case_ , ) lowercase =self._normalize_and_convert( snake_case_ , snake_case_ , input_points=snake_case_ , input_labels=snake_case_ , input_boxes=snake_case_ , return_tensors=snake_case_ , ) return encoding_image_processor def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=None , snake_case_=None , snake_case_="pt" , ): if input_points is not None: if len(snake_case_ ) != len(snake_case_ ): lowercase =[ self._normalize_coordinates(self.target_size , snake_case_ , original_sizes[0] ) for point in input_points ] else: lowercase =[ self._normalize_coordinates(self.target_size , snake_case_ , snake_case_ ) for point, original_size in zip(snake_case_ , snake_case_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: lowercase , lowercase =self._pad_points_and_labels(snake_case_ , snake_case_ ) lowercase =np.array(snake_case_ ) if input_labels is not None: lowercase =np.array(snake_case_ ) if input_boxes is not None: if len(snake_case_ ) != len(snake_case_ ): lowercase =[ self._normalize_coordinates(self.target_size , snake_case_ , original_sizes[0] , is_bounding_box=snake_case_ ) for box in input_boxes ] else: lowercase =[ self._normalize_coordinates(self.target_size , snake_case_ , snake_case_ , is_bounding_box=snake_case_ ) for box, original_size in zip(snake_case_ , snake_case_ ) ] lowercase =np.array(snake_case_ ) if input_boxes is not None: if return_tensors == "pt": lowercase =torch.from_numpy(snake_case_ ) # boxes batch size of 1 by default lowercase =input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": lowercase =tf.convert_to_tensor(snake_case_ ) # boxes batch size of 1 by default lowercase =tf.expand_dims(snake_case_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'''input_boxes''': input_boxes} ) if input_points is not None: if return_tensors == "pt": lowercase =torch.from_numpy(snake_case_ ) # point batch size of 1 by default lowercase =input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": lowercase =tf.convert_to_tensor(snake_case_ ) # point batch size of 1 by default lowercase =tf.expand_dims(snake_case_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'''input_points''': input_points} ) if input_labels is not None: if return_tensors == "pt": lowercase =torch.from_numpy(snake_case_ ) # point batch size of 1 by default lowercase =input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": lowercase =tf.convert_to_tensor(snake_case_ ) # point batch size of 1 by default lowercase =tf.expand_dims(snake_case_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'''input_labels''': input_labels} ) return encoding_image_processor def _A( self , snake_case_ , snake_case_ ): lowercase =max([point.shape[0] for point in input_points] ) lowercase =[] for i, point in enumerate(snake_case_ ): if point.shape[0] != expected_nb_points: lowercase =np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) lowercase =np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(snake_case_ ) lowercase =processed_input_points return input_points, input_labels def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=False ): lowercase , lowercase =original_size lowercase , lowercase =self.image_processor._get_preprocess_shape(snake_case_ , longest_edge=snake_case_ ) lowercase =deepcopy(snake_case_ ).astype(snake_case_ ) if is_bounding_box: lowercase =coords.reshape(-1 , 2 , 2 ) lowercase =coords[..., 0] * (new_w / old_w) lowercase =coords[..., 1] * (new_h / old_h) if is_bounding_box: lowercase =coords.reshape(-1 , 4 ) return coords def _A( self , snake_case_=None , snake_case_=None , snake_case_=None , ): if input_points is not None: if hasattr(snake_case_ , '''numpy''' ): # Checks for TF or Torch tensor lowercase =input_points.numpy().tolist() if not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_points[0] , snake_case_ ): raise ValueError('''Input points must be a list of list of floating points.''' ) lowercase =[np.array(snake_case_ ) for input_point in input_points] else: lowercase =None if input_labels is not None: if hasattr(snake_case_ , '''numpy''' ): lowercase =input_labels.numpy().tolist() if not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_labels[0] , snake_case_ ): raise ValueError('''Input labels must be a list of list integers.''' ) lowercase =[np.array(snake_case_ ) for label in input_labels] else: lowercase =None if input_boxes is not None: if hasattr(snake_case_ , '''numpy''' ): lowercase =input_boxes.numpy().tolist() if ( not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_boxes[0] , snake_case_ ) or not isinstance(input_boxes[0][0] , snake_case_ ) ): raise ValueError('''Input boxes must be a list of list of list of floating points.''' ) lowercase =[np.array(snake_case_ ).astype(np.floataa ) for box in input_boxes] else: lowercase =None return input_points, input_labels, input_boxes @property def _A( self ): lowercase =self.image_processor.model_input_names return list(dict.fromkeys(snake_case_ ) ) def _A( self , *snake_case_ , **snake_case_ ): return self.image_processor.post_process_masks(*snake_case_ , **snake_case_ )
72
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __magic_name__ ( unittest.TestCase ): def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=18 , snake_case_=30 , snake_case_=4_00 , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=None , ): lowercase =size if size is not None else {'''shortest_edge''': 20} lowercase =crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} lowercase =parent lowercase =batch_size lowercase =num_channels lowercase =image_size lowercase =min_resolution lowercase =max_resolution lowercase =do_resize lowercase =size lowercase =do_center_crop lowercase =crop_size def _A( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = MobileNetVaImageProcessor if is_vision_available() else None def _A( self ): lowercase =MobileNetVaImageProcessingTester(self ) @property def _A( self ): return self.image_processor_tester.prepare_image_processor_dict() def _A( self ): lowercase =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , '''do_resize''' ) ) self.assertTrue(hasattr(snake_case_ , '''size''' ) ) self.assertTrue(hasattr(snake_case_ , '''do_center_crop''' ) ) self.assertTrue(hasattr(snake_case_ , '''crop_size''' ) ) def _A( self ): lowercase =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 20} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) lowercase =self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def _A( self ): pass def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
72
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
1
'''simple docstring''' import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def UpperCamelCase ( lowercase_ : Dict , lowercase_ : List[Any] ) -> Any: '''simple docstring''' assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def UpperCamelCase ( lowercase_ : str , lowercase_ : int , lowercase_ : Optional[Any] ) -> Tuple: '''simple docstring''' lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase =ParquetDatasetReader(lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_parquet_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : str , lowercase_ : Any ) -> int: '''simple docstring''' lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase =features.copy() if features else default_expected_features lowercase =( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) lowercase =ParquetDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : List[str] , lowercase_ : List[Any] ) -> Optional[int]: '''simple docstring''' lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase =ParquetDatasetReader(lowercase_ , cache_dir=lowercase_ , split=lowercase_ ).read() _check_parquet_dataset(lowercase_ , lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : Tuple , lowercase_ : Optional[Any] ) -> Any: '''simple docstring''' if issubclass(lowercase_ , lowercase_ ): lowercase =parquet_path elif issubclass(lowercase_ , lowercase_ ): lowercase =[parquet_path] lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase =ParquetDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_ , lowercase_ ) def UpperCamelCase ( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Any=("train",) ) -> Optional[Any]: '''simple docstring''' assert isinstance(lowercase_ , lowercase_ ) for split in splits: lowercase =dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def UpperCamelCase ( lowercase_ : Dict , lowercase_ : Dict , lowercase_ : Optional[int] ) -> int: '''simple docstring''' lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase =ParquetDatasetReader( {'''train''': parquet_path} , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_parquet_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : List[Any] , lowercase_ : Optional[int] ) -> List[str]: '''simple docstring''' lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase =features.copy() if features else default_expected_features lowercase =( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) lowercase =ParquetDatasetReader({'''train''': parquet_path} , features=lowercase_ , cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Tuple ) -> Optional[int]: '''simple docstring''' if split: lowercase ={split: parquet_path} else: lowercase ='''train''' lowercase ={'''train''': parquet_path, '''test''': parquet_path} lowercase =tmp_path / '''cache''' lowercase ={'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase =ParquetDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_ , lowercase_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' lowercase =ParquetDatasetWriter(lowercase_ , tmp_path / '''foo.parquet''' ) assert writer.write() > 0 lowercase =pq.ParquetFile(tmp_path / '''foo.parquet''' ) lowercase =pf.read() assert dataset.data.table == output_table def UpperCamelCase ( lowercase_ : Tuple , lowercase_ : List[str] ) -> str: '''simple docstring''' lowercase =str(shared_datadir / '''test_image_rgb.jpg''' ) lowercase ={'''image''': [image_path]} lowercase =Features({'''image''': Image()} ) lowercase =Dataset.from_dict(lowercase_ , features=lowercase_ ) lowercase =ParquetDatasetWriter(lowercase_ , tmp_path / '''foo.parquet''' ) assert writer.write() > 0 lowercase =Dataset.from_parquet(str(tmp_path / '''foo.parquet''' ) ) assert dataset.features == reloaded_dataset.features lowercase =ParquetDatasetReader(str(tmp_path / '''foo.parquet''' ) , streaming=lowercase_ ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( '''feature, expected''' , [ (Features({'''foo''': Value('''int32''' )} ), None), (Features({'''image''': Image(), '''foo''': Value('''int32''' )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({'''nested''': Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ] , ) def UpperCamelCase ( lowercase_ : str , lowercase_ : List[Any] ) -> List[Any]: '''simple docstring''' assert get_writer_batch_size(lowercase_ ) == expected
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split _UpperCAmelCase : List[Any] = datasets.load_iris() _UpperCAmelCase : Dict = np.array(data['''data''']) _UpperCAmelCase : Union[str, Any] = np.array(data['''target''']) _UpperCAmelCase : int = data['''target_names'''] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = train_test_split(X, y) def UpperCamelCase ( lowercase_ : str , lowercase_ : Optional[Any] ) -> int: '''simple docstring''' return np.linalg.norm(np.array(lowercase_ ) - np.array(lowercase_ ) ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : Any , lowercase_ : int , lowercase_ : Tuple , lowercase_ : Tuple=5 ) -> List[Any]: '''simple docstring''' lowercase =zip(lowercase_ , lowercase_ ) # List of distances of all points from the point to be classified lowercase =[] for data_point in data: lowercase =euclidean_distance(data_point[0] , lowercase_ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. lowercase =[i[1] for i in sorted(lowercase_ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified lowercase =Counter(lowercase_ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
72
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _UpperCAmelCase : Any = True from torch.cuda.amp import autocast _UpperCAmelCase : Any = logging.getLogger(__name__) def UpperCamelCase ( lowercase_ : Optional[int]=None , lowercase_ : str=None ) -> Union[str, Any]: '''simple docstring''' return field(default_factory=lambda: default , metadata=lowercase_ ) @dataclass class __magic_name__ : UpperCamelCase__ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase__ = field( default=0.1 , metadata={'help': 'The dropout ratio for the attention probabilities.'} ) UpperCamelCase__ = field( default=0.1 , metadata={'help': 'The dropout ratio for activations inside the fully connected layer.'} ) UpperCamelCase__ = field( default=0.1 , metadata={ 'help': 'The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.' } , ) UpperCamelCase__ = field( default=0.1 , metadata={'help': 'The dropout probabilitiy for all 1D convolutional layers in feature extractor.'} , ) UpperCamelCase__ = field( default=0.0_5 , metadata={ 'help': ( 'Propability of each feature vector along the time axis to be chosen as the start of the vector' 'span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature' 'vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.' ) } , ) UpperCamelCase__ = field(default=0.0 , metadata={'help': 'The LayerDrop probability.'} ) @dataclass class __magic_name__ : UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase__ = field( default='train+validation' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of validation examples to this ' 'value if set.' ) } , ) UpperCamelCase__ = list_field( default=[',', '?', '.', '!', '-', ';', ':', '""', '%', '\'', '"', '�'] , metadata={'help': 'A list of characters to remove from the transcripts.'} , ) @dataclass class __magic_name__ : UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None def __call__( self , snake_case_ ): # split inputs and labels since they have to be of different lenghts and need # different padding methods lowercase =[{'''input_values''': feature['''input_values''']} for feature in features] lowercase =[{'''input_ids''': feature['''labels''']} for feature in features] lowercase =self.processor.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) lowercase =self.processor.pad( labels=snake_case_ , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='''pt''' , ) # replace padding with -100 to ignore loss correctly lowercase =labels_batch['''input_ids'''].masked_fill(labels_batch.attention_mask.ne(1 ) , -1_00 ) lowercase =labels return batch class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def _A( self , snake_case_ , snake_case_ ): model.train() lowercase =self._prepare_inputs(snake_case_ ) if self.use_amp: with autocast(): lowercase =self.compute_loss(snake_case_ , snake_case_ ) else: lowercase =self.compute_loss(snake_case_ , snake_case_ ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": lowercase =loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowercase =loss.sum() / (inputs['''labels'''] >= 0).sum() else: raise ValueError(f'{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']' ) if self.args.gradient_accumulation_steps > 1: lowercase =loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(snake_case_ ).backward() elif self.use_apex: with amp.scale_loss(snake_case_ , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(snake_case_ ) else: loss.backward() return loss.detach() def UpperCamelCase ( ) -> Tuple: '''simple docstring''' lowercase =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowercase , lowercase , lowercase =parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowercase , lowercase , lowercase =parser.parse_args_into_dataclasses() # Detecting last checkpoint. lowercase =None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase =get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('''Training/evaluation parameters %s''' , lowercase_ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: lowercase =datasets.load_dataset( '''common_voice''' , data_args.dataset_config_name , split=data_args.train_split_name ) lowercase =datasets.load_dataset('''common_voice''' , data_args.dataset_config_name , split='''test''' ) # Create and save tokenizer lowercase =f'[{"".join(data_args.chars_to_ignore )}]' def remove_special_characters(lowercase_ : List[Any] ): lowercase =re.sub(lowercase_ , '''''' , batch['''sentence'''] ).lower() + ''' ''' return batch lowercase =train_dataset.map(lowercase_ , remove_columns=['''sentence'''] ) lowercase =eval_dataset.map(lowercase_ , remove_columns=['''sentence'''] ) def extract_all_chars(lowercase_ : Optional[int] ): lowercase =''' '''.join(batch['''text'''] ) lowercase =list(set(lowercase_ ) ) return {"vocab": [vocab], "all_text": [all_text]} lowercase =train_dataset.map( lowercase_ , batched=lowercase_ , batch_size=-1 , keep_in_memory=lowercase_ , remove_columns=train_dataset.column_names , ) lowercase =train_dataset.map( lowercase_ , batched=lowercase_ , batch_size=-1 , keep_in_memory=lowercase_ , remove_columns=eval_dataset.column_names , ) lowercase =list(set(vocab_train['''vocab'''][0] ) | set(vocab_test['''vocab'''][0] ) ) lowercase ={v: k for k, v in enumerate(lowercase_ )} lowercase =vocab_dict[''' '''] del vocab_dict[" "] lowercase =len(lowercase_ ) lowercase =len(lowercase_ ) with open('''vocab.json''' , '''w''' ) as vocab_file: json.dump(lowercase_ , lowercase_ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase =WavaVecaCTCTokenizer( '''vocab.json''' , unk_token='''[UNK]''' , pad_token='''[PAD]''' , word_delimiter_token='''|''' , ) lowercase =WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0.0 , do_normalize=lowercase_ , return_attention_mask=lowercase_ ) lowercase =WavaVecaProcessor(feature_extractor=lowercase_ , tokenizer=lowercase_ ) lowercase =WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='''mean''' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: lowercase =min(len(lowercase_ ) , data_args.max_train_samples ) lowercase =train_dataset.select(range(lowercase_ ) ) if data_args.max_val_samples is not None: lowercase =eval_dataset.select(range(data_args.max_val_samples ) ) lowercase =torchaudio.transforms.Resample(4_8_0_0_0 , 1_6_0_0_0 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(lowercase_ : Optional[int] ): lowercase , lowercase =torchaudio.load(batch['''path'''] ) lowercase =resampler(lowercase_ ).squeeze().numpy() lowercase =1_6_0_0_0 lowercase =batch['''text'''] return batch lowercase =train_dataset.map( lowercase_ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) lowercase =eval_dataset.map( lowercase_ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(lowercase_ : int ): # check that all files have the correct sampling rate assert ( len(set(batch['''sampling_rate'''] ) ) == 1 ), f'Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.' lowercase =processor( audio=batch['''speech'''] , text=batch['''target_text'''] , sampling_rate=batch['''sampling_rate'''][0] ) batch.update(lowercase_ ) return batch lowercase =train_dataset.map( lowercase_ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowercase_ , num_proc=data_args.preprocessing_num_workers , ) lowercase =eval_dataset.map( lowercase_ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowercase_ , num_proc=data_args.preprocessing_num_workers , ) # Metric lowercase =datasets.load_metric('''wer''' ) def compute_metrics(lowercase_ : List[str] ): lowercase =pred.predictions lowercase =np.argmax(lowercase_ , axis=-1 ) lowercase =processor.tokenizer.pad_token_id lowercase =processor.batch_decode(lowercase_ ) # we do not want to group tokens when computing the metrics lowercase =processor.batch_decode(pred.label_ids , group_tokens=lowercase_ ) lowercase =wer_metric.compute(predictions=lowercase_ , references=lowercase_ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator lowercase =DataCollatorCTCWithPadding(processor=lowercase_ , padding=lowercase_ ) # Initialize our Trainer lowercase =CTCTrainer( model=lowercase_ , data_collator=lowercase_ , args=lowercase_ , compute_metrics=lowercase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: lowercase =last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): lowercase =model_args.model_name_or_path else: lowercase =None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) lowercase =trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() lowercase =train_result.metrics lowercase =( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase_ ) ) lowercase =min(lowercase_ , len(lowercase_ ) ) trainer.log_metrics('''train''' , lowercase_ ) trainer.save_metrics('''train''' , lowercase_ ) trainer.save_state() # Evaluation lowercase ={} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase =trainer.evaluate() lowercase =data_args.max_val_samples if data_args.max_val_samples is not None else len(lowercase_ ) lowercase =min(lowercase_ , len(lowercase_ ) ) trainer.log_metrics('''eval''' , lowercase_ ) trainer.save_metrics('''eval''' , lowercase_ ) return results if __name__ == "__main__": main()
72
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
1
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=2 , snake_case_=24 , snake_case_=16 , snake_case_=True , snake_case_=True , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=10 , snake_case_=0.02 , snake_case_=None , snake_case_=2 , snake_case_=2 , ): lowercase =parent lowercase =batch_size lowercase =patch_size lowercase =max_length lowercase =num_mel_bins lowercase =is_training lowercase =use_labels lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =type_sequence_label_size lowercase =initializer_range lowercase =scope lowercase =frequency_stride lowercase =time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowercase =(self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowercase =(self.max_length - self.patch_size) // self.time_stride + 1 lowercase =frequency_out_dimension * time_out_dimension lowercase =num_patches + 2 def _A( self ): lowercase =floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowercase =None if self.use_labels: lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase =self.get_config() return config, input_values, labels def _A( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =ASTModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={'''input_values''': input_values} return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) UpperCamelCase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def _A( self ): lowercase =ASTModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def _A( self ): pass def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase =model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =['''input_values'''] self.assertListEqual(arg_names[:1] , snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =ASTModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def UpperCamelCase ( ) -> Optional[Any]: '''simple docstring''' lowercase =hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) lowercase , lowercase =torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class __magic_name__ ( unittest.TestCase ): @cached_property def _A( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def _A( self ): lowercase =self.default_feature_extractor lowercase =ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(snake_case_ ) lowercase =self.default_feature_extractor lowercase , lowercase =prepare_audio() lowercase =audio.squeeze().numpy() lowercase =feature_extractor(snake_case_ , sampling_rate=snake_case_ , return_tensors='''pt''' ).to(snake_case_ ) # forward pass with torch.no_grad(): lowercase =model(**snake_case_ ) # verify the logits lowercase =torch.Size((1, 5_27) ) self.assertEqual(outputs.logits.shape , snake_case_ ) lowercase =torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1E-4 ) )
72
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
1
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
1
'''simple docstring''' from statistics import mean import numpy as np def UpperCamelCase ( lowercase_ : list , lowercase_ : list , lowercase_ : list , lowercase_ : int ) -> list: '''simple docstring''' lowercase =0 # Number of processes finished lowercase =0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. lowercase =[0] * no_of_process # List to include calculation results lowercase =[0] * no_of_process # Sort by arrival time. lowercase =[burst_time[i] for i in np.argsort(lowercase_ )] lowercase =[process_name[i] for i in np.argsort(lowercase_ )] arrival_time.sort() while no_of_process > finished_process_count: lowercase =0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: lowercase =arrival_time[i] lowercase =0 # Index showing the location of the process being performed lowercase =0 # Saves the current response ratio. lowercase =0 for i in range(0 , lowercase_ ): if finished_process[i] == 0 and arrival_time[i] <= current_time: lowercase =(burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: lowercase =temp lowercase =i # Calculate the turn around time lowercase =current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. lowercase =1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def UpperCamelCase ( lowercase_ : list , lowercase_ : list , lowercase_ : list , lowercase_ : int ) -> list: '''simple docstring''' lowercase =[0] * no_of_process for i in range(0 , lowercase_ ): lowercase =turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": _UpperCAmelCase : str = 5 _UpperCAmelCase : Optional[Any] = ['''A''', '''B''', '''C''', '''D''', '''E'''] _UpperCAmelCase : Dict = [1, 2, 3, 4, 5] _UpperCAmelCase : List[Any] = [1, 2, 3, 4, 5] _UpperCAmelCase : Optional[Any] = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) _UpperCAmelCase : Optional[Any] = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print('''Process name \tArrival time \tBurst time \tTurn around time \tWaiting time''') for i in range(0, no_of_process): print( F"""{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t""" F"""{turn_around_time[i]}\t\t\t{waiting_time[i]}""" ) print(F"""average waiting time : {mean(waiting_time):.5f}""") print(F"""average turn around time : {mean(turn_around_time):.5f}""")
72
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
1
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
1
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ ): lowercase =False super().__init__(snake_case_ , snake_case_ ) lowercase =self.image_processor def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None: lowercase =self.tokenizer lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) return text_encoding # add pixel_values lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) else: lowercase =None if text_encoding is not None: encoding_image_processor.update(snake_case_ ) return encoding_image_processor def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
72
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule _UpperCAmelCase : Optional[int] = {'''tokenization_byt5''': ['''ByT5Tokenizer''']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _UpperCAmelCase : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
1
'''simple docstring''' import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def _A( self ): lowercase =pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def _A( self ): with self.assertRaises(snake_case_ ): lowercase =pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def _A( self ): with self.assertRaises(snake_case_ ): lowercase =pa.array(TypedSequence([1, 2, 3] , try_type=Value('''bool''' ) , type=Value('''int64''' ) ) ) def _A( self ): lowercase =pa.array(TypedSequence([1, 2, 3] , type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _A( self ): with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): lowercase =pa.array(TypedSequence(['''foo''', '''bar'''] , type=Value('''int64''' ) ) ) def _A( self ): lowercase =pa.array(TypedSequence([1, 2, 3] , try_type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _A( self ): lowercase =pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=Value('''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) def _A( self ): lowercase =pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def _A( self ): with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): lowercase =pa.array(TypedSequence(['''foo''', '''bar'''] , type=ArrayaD((1, 3) , '''int64''' ) ) ) def _A( self ): lowercase =pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def _A( self ): lowercase =pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def _A( self ): import PIL.Image lowercase =PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( '''datasets.arrow_writer.cast_to_python_objects''' , side_effect=snake_case_ ) as mock_cast_to_python_objects: lowercase =pa.array(TypedSequence([{'''path''': None, '''bytes''': B'''image_bytes'''}, pil_image] , type=Image() ) ) lowercase , lowercase =mock_cast_to_python_objects.call_args_list[-1] self.assertIn('''optimize_list_casting''' , snake_case_ ) self.assertFalse(kwargs['''optimize_list_casting'''] ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : int ) -> Optional[int]: '''simple docstring''' lowercase =pa.BufferReader(lowercase_ ) if isinstance(lowercase_ , pa.Buffer ) else pa.memory_map(lowercase_ ) lowercase =pa.ipc.open_stream(lowercase_ ) lowercase =f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' lowercase =pa.BufferOutputStream() lowercase =pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: lowercase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def UpperCamelCase ( ) -> int: '''simple docstring''' lowercase =pa.BufferOutputStream() lowercase =Features({'''labels''': ClassLabel(names=['''neg''', '''pos'''] )} ) with ArrowWriter(stream=lowercase_ , features=lowercase_ ) as writer: writer.write({'''labels''': 0} ) writer.write({'''labels''': 1} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata lowercase =pa.BufferReader(output.getvalue() ) lowercase =pa.ipc.open_stream(lowercase_ ) lowercase =f.read_all() lowercase =pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(lowercase_ ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) def UpperCamelCase ( lowercase_ : List[str] ) -> str: '''simple docstring''' lowercase =pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt='''split_name''' , check_duplicates=lowercase_ , ) as writer: with pytest.raises(lowercase_ ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=[1, 2] ) lowercase , lowercase =writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 1_0] ) def UpperCamelCase ( lowercase_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' lowercase =pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt='''split_name''' , check_duplicates=lowercase_ , ) as writer: with pytest.raises(lowercase_ ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=1_0 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=1_0 ) lowercase , lowercase =writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 1_0] ) def UpperCamelCase ( lowercase_ : List[Any] ) -> Any: '''simple docstring''' lowercase =pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt='''split_name''' , check_duplicates=lowercase_ , ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=1 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=2 ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : Union[str, Any] ) -> Any: '''simple docstring''' lowercase =pa.BufferOutputStream() lowercase =pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) writer.write_batch({'''col_1''': [], '''col_2''': []} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: lowercase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def UpperCamelCase ( lowercase_ : str , lowercase_ : List[str] ) -> int: '''simple docstring''' lowercase =pa.BufferOutputStream() lowercase =pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_table(pa.Table.from_pydict({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: lowercase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : Tuple ) -> int: '''simple docstring''' lowercase =pa.BufferOutputStream() lowercase =pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_row(pa.Table.from_pydict({'''col_1''': ['''foo'''], '''col_2''': [1]} ) ) writer.write_row(pa.Table.from_pydict({'''col_1''': ['''bar'''], '''col_2''': [2]} ) ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: lowercase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def UpperCamelCase ( ) -> Union[str, Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: lowercase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} lowercase =os.path.join(lowercase_ , '''test.arrow''' ) with ArrowWriter(path=lowercase_ , schema=pa.schema(lowercase_ ) ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(lowercase_ , 1 ) def UpperCamelCase ( lowercase_ : List[str] ) -> Dict: '''simple docstring''' if pa.types.is_list(lowercase_ ): return get_base_dtype(arr_type.value_type ) else: return arr_type def UpperCamelCase ( lowercase_ : Dict , lowercase_ : str ) -> Optional[int]: '''simple docstring''' if isinstance(lst[0] , lowercase_ ): change_first_primitive_element_in_list(lst[0] , lowercase_ ) else: lowercase =value @pytest.mark.parametrize('''optimized_int_type, expected_dtype''' , [(None, pa.intaa()), (Value('''int32''' ), pa.intaa())] ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def UpperCamelCase ( lowercase_ : str , lowercase_ : List[Any] , lowercase_ : Optional[int] ) -> Optional[Any]: '''simple docstring''' lowercase =pa.array(TypedSequence(lowercase_ , optimized_int_type=lowercase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( '''col, expected_dtype''' , [ ('''attention_mask''', pa.inta()), ('''special_tokens_mask''', pa.inta()), ('''token_type_ids''', pa.inta()), ('''input_ids''', pa.intaa()), ('''other''', pa.intaa()), ] , ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def UpperCamelCase ( lowercase_ : Dict , lowercase_ : List[Any] , lowercase_ : Union[str, Any] ) -> Any: '''simple docstring''' lowercase =pa.array(OptimizedTypedSequence(lowercase_ , col=lowercase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications lowercase =copy.deepcopy(lowercase_ ) lowercase =np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(lowercase_ , lowercase_ ) lowercase =pa.array(OptimizedTypedSequence(lowercase_ , col=lowercase_ ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize('''raise_exception''' , [False, True] ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : Tuple ) -> Optional[int]: '''simple docstring''' lowercase =str(tmp_path / '''dataset-train.arrow''' ) try: with ArrowWriter(path=lowercase_ ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def UpperCamelCase ( lowercase_ : List[Any] ) -> str: '''simple docstring''' lowercase ='''mock://dataset-train.arrow''' with ArrowWriter(path=lowercase_ , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(lowercase_ ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(lowercase_ ) def UpperCamelCase ( ) -> Optional[int]: '''simple docstring''' lowercase =pa.BufferOutputStream() with ParquetWriter(stream=lowercase_ ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) lowercase , lowercase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 lowercase =pa.BufferReader(output.getvalue() ) lowercase =pq.read_table(lowercase_ ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize('''embed_local_files''' , [False, True] ) def UpperCamelCase ( lowercase_ : Dict , lowercase_ : Any ) -> str: '''simple docstring''' import PIL.Image lowercase =str(tmp_path / '''test_image_rgb.jpg''' ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(lowercase_ , format='''png''' ) lowercase =pa.BufferOutputStream() with ParquetWriter( stream=lowercase_ , features=Features({'''image''': Image()} ) , embed_local_files=lowercase_ ) as writer: writer.write({'''image''': image_path} ) writer.finalize() lowercase =pa.BufferReader(output.getvalue() ) lowercase =pq.read_table(lowercase_ ) lowercase =pa_table.to_pydict() if embed_local_files: assert isinstance(out['''image'''][0]['''path'''] , lowercase_ ) with open(lowercase_ , '''rb''' ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def UpperCamelCase ( ) -> str: '''simple docstring''' lowercase =pa.schema([pa.field('''col_1''' , pa.string() , nullable=lowercase_ )] ) lowercase =pa.BufferOutputStream() with ArrowWriter(stream=lowercase_ ) as writer: writer._build_writer(inferred_schema=lowercase_ ) assert writer._schema == pa.schema([pa.field('''col_1''' , pa.string() )] )
72
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
1
'''simple docstring''' from __future__ import annotations def UpperCamelCase ( lowercase_ : list[int | str] ) -> None: '''simple docstring''' create_state_space_tree(lowercase_ , [] , 0 , [0 for i in range(len(lowercase_ ) )] ) def UpperCamelCase ( lowercase_ : list[int | str] , lowercase_ : list[int | str] , lowercase_ : int , lowercase_ : list[int] , ) -> None: '''simple docstring''' if index == len(lowercase_ ): print(lowercase_ ) return for i in range(len(lowercase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) lowercase =True create_state_space_tree(lowercase_ , lowercase_ , index + 1 , lowercase_ ) current_sequence.pop() lowercase =False _UpperCAmelCase : list[int | str] = [3, 1, 2, 4] generate_all_permutations(sequence) _UpperCAmelCase : list[int | str] = ["A", "B", "C"] generate_all_permutations(sequence_a)
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
1
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
1
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 _UpperCAmelCase : List[Any] = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') _UpperCAmelCase : List[str] = get_tests_dir('''fixtures/vocab.json''') _UpperCAmelCase : Optional[int] = get_tests_dir('''fixtures''') class __magic_name__ ( unittest.TestCase ): UpperCamelCase__ = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] def _A( self ): lowercase =0 def _A( self ): lowercase =AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): with tempfile.TemporaryDirectory() as tmpdirname: lowercase =WavaVecaConfig() lowercase =AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) # save in new folder model_config.save_pretrained(snake_case_ ) processor.save_pretrained(snake_case_ ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(snake_case_ , os.path.join(snake_case_ , snake_case_ ) ) copyfile(snake_case_ , os.path.join(snake_case_ , '''vocab.json''' ) ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): with tempfile.TemporaryDirectory() as tmpdirname: lowercase =WavaVecaFeatureExtractor() lowercase =AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) lowercase =WavaVecaProcessor(snake_case_ , snake_case_ ) # save in new folder processor.save_pretrained(snake_case_ ) # drop `processor_class` in tokenizer with open(os.path.join(snake_case_ , snake_case_ ) , '''r''' ) as f: lowercase =json.load(snake_case_ ) config_dict.pop('''processor_class''' ) with open(os.path.join(snake_case_ , snake_case_ ) , '''w''' ) as f: f.write(json.dumps(snake_case_ ) ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): with tempfile.TemporaryDirectory() as tmpdirname: lowercase =WavaVecaFeatureExtractor() lowercase =AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) lowercase =WavaVecaProcessor(snake_case_ , snake_case_ ) # save in new folder processor.save_pretrained(snake_case_ ) # drop `processor_class` in feature extractor with open(os.path.join(snake_case_ , snake_case_ ) , '''r''' ) as f: lowercase =json.load(snake_case_ ) config_dict.pop('''processor_class''' ) with open(os.path.join(snake_case_ , snake_case_ ) , '''w''' ) as f: f.write(json.dumps(snake_case_ ) ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): with tempfile.TemporaryDirectory() as tmpdirname: lowercase =WavaVecaConfig(processor_class='''Wav2Vec2Processor''' ) model_config.save_pretrained(snake_case_ ) # copy relevant files copyfile(snake_case_ , os.path.join(snake_case_ , '''vocab.json''' ) ) # create emtpy sample processor with open(os.path.join(snake_case_ , snake_case_ ) , '''w''' ) as f: f.write('''{}''' ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def _A( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(snake_case_ ): lowercase =AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(snake_case_ ): lowercase =AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=snake_case_ ) lowercase =AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=snake_case_ ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) lowercase =processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , '''NewFeatureExtractor''' ) lowercase =processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) # Test we can also load the slow version lowercase =AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=snake_case_ , use_fast=snake_case_ ) lowercase =new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , '''NewTokenizer''' ) else: self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) def _A( self ): try: AutoConfig.register('''custom''' , snake_case_ ) AutoFeatureExtractor.register(snake_case_ , snake_case_ ) AutoTokenizer.register(snake_case_ , slow_tokenizer_class=snake_case_ ) AutoProcessor.register(snake_case_ , snake_case_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(snake_case_ ): AutoProcessor.register(snake_case_ , snake_case_ ) # Now that the config is registered, it can be used as any other config with the auto-API lowercase =CustomFeatureExtractor.from_pretrained(snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: lowercase =os.path.join(snake_case_ , '''vocab.txt''' ) with open(snake_case_ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) lowercase =CustomTokenizer(snake_case_ ) lowercase =CustomProcessor(snake_case_ , snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(snake_case_ ) lowercase =AutoProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _A( self ): class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = False class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = False class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'AutoFeatureExtractor' UpperCamelCase__ = 'AutoTokenizer' UpperCamelCase__ = False try: AutoConfig.register('''custom''' , snake_case_ ) AutoFeatureExtractor.register(snake_case_ , snake_case_ ) AutoTokenizer.register(snake_case_ , slow_tokenizer_class=snake_case_ ) AutoProcessor.register(snake_case_ , snake_case_ ) # If remote code is not set, the default is to use local classes. lowercase =AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. lowercase =AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=snake_case_ ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. lowercase =AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=snake_case_ ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _A( self ): lowercase =AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(processor.__class__.__name__ , '''BertTokenizerFast''' ) def _A( self ): lowercase =AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-convnext''' ) self.assertEqual(processor.__class__.__name__ , '''ConvNextImageProcessor''' ) @is_staging_test class __magic_name__ ( unittest.TestCase ): UpperCamelCase__ = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def _A( cls ): lowercase =TOKEN HfFolder.save_token(snake_case_ ) @classmethod def _A( cls ): try: delete_repo(token=cls._token , repo_id='''test-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-processor''' ) except HTTPError: pass def _A( self ): lowercase =WavaVecaProcessor.from_pretrained(snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(snake_case_ , '''test-processor''' ) , push_to_hub=snake_case_ , use_auth_token=self._token ) lowercase =WavaVecaProcessor.from_pretrained(f'{USER}/test-processor' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(snake_case_ , getattr(new_processor.feature_extractor , snake_case_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _A( self ): lowercase =WavaVecaProcessor.from_pretrained(snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(snake_case_ , '''test-processor-org''' ) , push_to_hub=snake_case_ , use_auth_token=self._token , organization='''valid_org''' , ) lowercase =WavaVecaProcessor.from_pretrained('''valid_org/test-processor-org''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(snake_case_ , getattr(new_processor.feature_extractor , snake_case_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _A( self ): CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() lowercase =CustomFeatureExtractor.from_pretrained(snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: lowercase =os.path.join(snake_case_ , '''vocab.txt''' ) with open(snake_case_ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) lowercase =CustomTokenizer(snake_case_ ) lowercase =CustomProcessor(snake_case_ , snake_case_ ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f'{USER}/test-dynamic-processor' , token=self._token ) lowercase =Repository(snake_case_ , clone_from=f'{USER}/test-dynamic-processor' , token=self._token ) processor.save_pretrained(snake_case_ ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { '''AutoFeatureExtractor''': '''custom_feature_extraction.CustomFeatureExtractor''', '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(snake_case_ , '''tokenizer_config.json''' ) ) as f: lowercase =json.load(snake_case_ ) self.assertDictEqual( tokenizer_config['''auto_map'''] , { '''AutoTokenizer''': ['''custom_tokenization.CustomTokenizer''', None], '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(snake_case_ , '''custom_feature_extraction.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(snake_case_ , '''custom_tokenization.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(snake_case_ , '''custom_processing.py''' ) ) ) repo.push_to_hub() lowercase =AutoProcessor.from_pretrained(f'{USER}/test-dynamic-processor' , trust_remote_code=snake_case_ ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , '''CustomProcessor''' )
72
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'facebook/bart-large-mnli' UpperCamelCase__ = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) UpperCamelCase__ = 'text_classifier' UpperCamelCase__ = AutoTokenizer UpperCamelCase__ = AutoModelForSequenceClassification UpperCamelCase__ = ['text', ['text']] UpperCamelCase__ = ['text'] def _A( self ): super().setup() lowercase =self.model.config lowercase =-1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): lowercase =int(snake_case_ ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def _A( self , snake_case_ , snake_case_ ): lowercase =labels return self.pre_processor( [text] * len(snake_case_ ) , [f'This example is {label}' for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def _A( self , snake_case_ ): lowercase =outputs.logits lowercase =torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
1
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
1
'''simple docstring''' import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 0 UpperCamelCase__ = False UpperCamelCase__ = 3.0 class __magic_name__ ( unittest.TestCase ): def _A( self ): # If no defaults are changed, `to_kwargs` returns an empty dict. self.assertDictEqual(MockClass().to_kwargs() , {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs() , {'''a''': 2} ) self.assertDictEqual(MockClass(a=2 , b=snake_case_ ).to_kwargs() , {'''a''': 2, '''b''': True} ) self.assertDictEqual(MockClass(a=2 , c=2.25 ).to_kwargs() , {'''a''': 2, '''c''': 2.25} ) @require_cuda def _A( self ): # If no defaults are changed, `to_kwargs` returns an empty dict. lowercase =GradScalerKwargs(init_scale=10_24 , growth_factor=2 ) AcceleratorState._reset_state() lowercase =Accelerator(mixed_precision='''fp16''' , kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) lowercase =accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale , 10_24.0 ) self.assertEqual(scaler._growth_factor , 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor , 0.5 ) self.assertEqual(scaler._growth_interval , 20_00 ) self.assertEqual(scaler._enabled , snake_case_ ) @require_multi_gpu def _A( self ): lowercase =['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__ )] execute_subprocess_async(snake_case_ , env=os.environ.copy() ) if __name__ == "__main__": _UpperCAmelCase : Union[str, Any] = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) _UpperCAmelCase : Dict = Accelerator(kwargs_handlers=[ddp_scaler]) _UpperCAmelCase : Tuple = torch.nn.Linear(1_00, 2_00) _UpperCAmelCase : Any = accelerator.prepare(model) # Check the values changed in kwargs _UpperCAmelCase : int = '''''' _UpperCAmelCase : str = model.bucket_bytes_cap // (10_24 * 10_24) if observed_bucket_cap_map != 15: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
72
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
1
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _UpperCAmelCase : int = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class __magic_name__ ( unittest.TestCase ): UpperCamelCase__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def _A( self ): lowercase =pipeline( task='''text-classification''' , model='''hf-internal-testing/tiny-random-distilbert''' , framework='''pt''' ) lowercase =text_classifier('''This is great !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}] ) lowercase =text_classifier('''This is great !''' , top_k=2 ) self.assertEqual( nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}] ) lowercase =text_classifier(['''This is great !''', '''This is bad'''] , top_k=2 ) self.assertEqual( nested_simplify(snake_case_ ) , [ [{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}], [{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}], ] , ) lowercase =text_classifier('''This is great !''' , top_k=1 ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}] ) # Legacy behavior lowercase =text_classifier('''This is great !''' , return_all_scores=snake_case_ ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}] ) lowercase =text_classifier('''This is great !''' , return_all_scores=snake_case_ ) self.assertEqual( nested_simplify(snake_case_ ) , [[{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}]] ) lowercase =text_classifier(['''This is great !''', '''Something else'''] , return_all_scores=snake_case_ ) self.assertEqual( nested_simplify(snake_case_ ) , [ [{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}], [{'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_1''', '''score''': 0.4_96}], ] , ) lowercase =text_classifier(['''This is great !''', '''Something else'''] , return_all_scores=snake_case_ ) self.assertEqual( nested_simplify(snake_case_ ) , [ {'''label''': '''LABEL_0''', '''score''': 0.5_04}, {'''label''': '''LABEL_0''', '''score''': 0.5_04}, ] , ) @require_torch def _A( self ): import torch lowercase =pipeline( task='''text-classification''' , model='''hf-internal-testing/tiny-random-distilbert''' , framework='''pt''' , device=torch.device('''cpu''' ) , ) lowercase =text_classifier('''This is great !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}] ) @require_tf def _A( self ): lowercase =pipeline( task='''text-classification''' , model='''hf-internal-testing/tiny-random-distilbert''' , framework='''tf''' ) lowercase =text_classifier('''This is great !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''LABEL_0''', '''score''': 0.5_04}] ) @slow @require_torch def _A( self ): lowercase =pipeline('''text-classification''' ) lowercase =text_classifier('''This is great !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''POSITIVE''', '''score''': 1.0}] ) lowercase =text_classifier('''This is bad !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''NEGATIVE''', '''score''': 1.0}] ) lowercase =text_classifier('''Birds are a type of animal''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''POSITIVE''', '''score''': 0.9_88}] ) @slow @require_tf def _A( self ): lowercase =pipeline('''text-classification''' , framework='''tf''' ) lowercase =text_classifier('''This is great !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''POSITIVE''', '''score''': 1.0}] ) lowercase =text_classifier('''This is bad !''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''NEGATIVE''', '''score''': 1.0}] ) lowercase =text_classifier('''Birds are a type of animal''' ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': '''POSITIVE''', '''score''': 0.9_88}] ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =TextClassificationPipeline(model=snake_case_ , tokenizer=snake_case_ ) return text_classifier, ["HuggingFace is in", "This is another test"] def _A( self , snake_case_ , snake_case_ ): lowercase =text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 lowercase ='''HuggingFace is in''' lowercase =text_classifier(snake_case_ ) self.assertEqual(nested_simplify(snake_case_ ) , [{'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}] ) self.assertTrue(outputs[0]['''label'''] in model.config.idalabel.values() ) lowercase =['''HuggingFace is in ''', '''Paris is in France'''] lowercase =text_classifier(snake_case_ ) self.assertEqual( nested_simplify(snake_case_ ) , [{'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}, {'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}] , ) self.assertTrue(outputs[0]['''label'''] in model.config.idalabel.values() ) self.assertTrue(outputs[1]['''label'''] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format lowercase =text_classifier(snake_case_ , top_k=snake_case_ ) lowercase =len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(snake_case_ ) , [[{'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}] * N, [{'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}] * N] , ) lowercase ={'''text''': '''HuggingFace is in ''', '''text_pair''': '''Paris is in France'''} lowercase =text_classifier(snake_case_ ) self.assertEqual( nested_simplify(snake_case_ ) , {'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )} , ) self.assertTrue(outputs['''label'''] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. lowercase =[['''HuggingFace is in ''', '''Paris is in France''']] with self.assertRaises(snake_case_ ): text_classifier(snake_case_ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility lowercase =text_classifier([[['''HuggingFace is in ''', '''Paris is in France''']]] ) self.assertEqual( nested_simplify(snake_case_ ) , [{'''label''': ANY(snake_case_ ), '''score''': ANY(snake_case_ )}] , ) self.assertTrue(outputs[0]['''label'''] in model.config.idalabel.values() )
72
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import math from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : str = { '''facebook/data2vec-base-960h''': '''https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json''', # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'data2vec-audio' def __init__( self , snake_case_=32 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_=30_72 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.02 , snake_case_=1E-5 , snake_case_="gelu" , snake_case_=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case_=(5, 2, 2, 2, 2, 2, 2) , snake_case_=(10, 3, 3, 3, 3, 2, 2) , snake_case_=False , snake_case_=16 , snake_case_=19 , snake_case_=5 , snake_case_=0.05 , snake_case_=10 , snake_case_=2 , snake_case_=0.0 , snake_case_=10 , snake_case_=0 , snake_case_="sum" , snake_case_=False , snake_case_=False , snake_case_=2_56 , snake_case_=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case_=(5, 3, 3, 1, 1) , snake_case_=(1, 2, 3, 1, 1) , snake_case_=5_12 , snake_case_=0 , snake_case_=1 , snake_case_=2 , snake_case_=False , snake_case_=3 , snake_case_=2 , snake_case_=3 , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ , pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ ) lowercase =hidden_size lowercase =feat_extract_activation lowercase =list(snake_case_ ) lowercase =list(snake_case_ ) lowercase =list(snake_case_ ) lowercase =conv_bias lowercase =num_conv_pos_embeddings lowercase =num_conv_pos_embedding_groups lowercase =conv_pos_kernel_size lowercase =len(self.conv_dim ) lowercase =num_hidden_layers lowercase =intermediate_size lowercase =hidden_act lowercase =num_attention_heads lowercase =hidden_dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =feat_proj_dropout lowercase =final_dropout lowercase =layerdrop lowercase =layer_norm_eps lowercase =initializer_range lowercase =vocab_size lowercase =use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,' f' `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase =mask_time_prob lowercase =mask_time_length lowercase =mask_time_min_masks lowercase =mask_feature_prob lowercase =mask_feature_length lowercase =mask_feature_min_masks # ctc loss lowercase =ctc_loss_reduction lowercase =ctc_zero_infinity # adapter lowercase =add_adapter lowercase =adapter_kernel_size lowercase =adapter_stride lowercase =num_adapter_layers lowercase =output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase =classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase =list(snake_case_ ) lowercase =list(snake_case_ ) lowercase =list(snake_case_ ) lowercase =xvector_output_dim @property def _A( self ): return math.prod(self.conv_stride )
72
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCAmelCase : Union[str, Any] = { '''configuration_mobilebert''': [ '''MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileBertConfig''', '''MobileBertOnnxConfig''', ], '''tokenization_mobilebert''': ['''MobileBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''MobileBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ '''MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileBertForMaskedLM''', '''MobileBertForMultipleChoice''', '''MobileBertForNextSentencePrediction''', '''MobileBertForPreTraining''', '''MobileBertForQuestionAnswering''', '''MobileBertForSequenceClassification''', '''MobileBertForTokenClassification''', '''MobileBertLayer''', '''MobileBertModel''', '''MobileBertPreTrainedModel''', '''load_tf_weights_in_mobilebert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ '''TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileBertForMaskedLM''', '''TFMobileBertForMultipleChoice''', '''TFMobileBertForNextSentencePrediction''', '''TFMobileBertForPreTraining''', '''TFMobileBertForQuestionAnswering''', '''TFMobileBertForSequenceClassification''', '''TFMobileBertForTokenClassification''', '''TFMobileBertMainLayer''', '''TFMobileBertModel''', '''TFMobileBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys _UpperCAmelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
1
'''simple docstring''' _UpperCAmelCase : dict[str, float] = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_02_17_66_34e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.35_5818, } def UpperCamelCase ( lowercase_ : str , lowercase_ : str , lowercase_ : float ) -> float: '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowercase =( f'Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n' f'Valid values are: {", ".join(lowercase_ )}' ) raise ValueError(lowercase_ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
72
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
1
'''simple docstring''' import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCamelCase ( lowercase_ : Any , lowercase_ : int ) -> Dict: '''simple docstring''' lowercase =torch.load(lowercase_ , map_location='''cpu''' ) lowercase =chkpt['''model'''] # We have the base model one level deeper than the original XLM repository lowercase ={} for k, v in state_dict.items(): if "pred_layer" in k: lowercase =v else: lowercase =v lowercase =chkpt['''params'''] lowercase ={n: v for n, v in config.items() if not isinstance(lowercase_ , (torch.FloatTensor, numpy.ndarray) )} lowercase =chkpt['''dico_word2id'''] lowercase ={s + '''</w>''' if s.find('''@@''' ) == -1 and i > 1_3 else s.replace('''@@''' , '''''' ): i for s, i in vocab.items()} # Save pytorch-model lowercase =pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME lowercase =pytorch_dump_folder_path + '''/''' + CONFIG_NAME lowercase =pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''vocab_file'''] print(f'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(lowercase_ , lowercase_ ) print(f'Save configuration file to {pytorch_config_dump_path}' ) with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowercase_ , indent=2 ) + '''\n''' ) print(f'Save vocab file to {pytorch_config_dump_path}' ) with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowercase_ , indent=2 ) + '''\n''' ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _UpperCAmelCase : List[Any] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
72
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
1
'''simple docstring''' # limitations under the License. from typing import Optional, Tuple, Union import torch from diffusers import DiffusionPipeline, ImagePipelineOutput class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , snake_case_ , snake_case_ ): super().__init__() self.register_modules(unet=snake_case_ , scheduler=snake_case_ ) @torch.no_grad() def __call__( self , snake_case_ = 1 , snake_case_ = None , snake_case_ = 50 , snake_case_ = "pil" , snake_case_ = True , **snake_case_ , ): lowercase =torch.randn( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=snake_case_ , ) lowercase =image.to(self.device ) # set step values self.scheduler.set_timesteps(snake_case_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowercase =self.unet(snake_case_ , snake_case_ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 lowercase =self.scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample lowercase =(image / 2 + 0.5).clamp(0 , 1 ) lowercase =image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowercase =self.numpy_to_pil(snake_case_ ) if not return_dict: return (image,), "This is a local test" return ImagePipelineOutput(images=snake_case_ ), "This is a local test"
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' import os def UpperCamelCase ( ) -> str: '''simple docstring''' with open(os.path.dirname(lowercase_ ) + '''/grid.txt''' ) as f: lowercase =[] # noqa: E741 for _ in range(2_0 ): l.append([int(lowercase_ ) for x in f.readline().split()] ) lowercase =0 # right for i in range(2_0 ): for j in range(1_7 ): lowercase =l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: lowercase =temp # down for i in range(1_7 ): for j in range(2_0 ): lowercase =l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: lowercase =temp # diagonal 1 for i in range(1_7 ): for j in range(1_7 ): lowercase =l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: lowercase =temp # diagonal 2 for i in range(1_7 ): for j in range(3 , 2_0 ): lowercase =l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: lowercase =temp return maximum if __name__ == "__main__": print(solution())
72
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : int = { '''vocab_file''': '''vocab.json''', '''tokenizer_config_file''': '''tokenizer_config.json''', '''merges_file''': '''merges.txt''', } _UpperCAmelCase : int = { '''vocab_file''': { '''facebook/s2t-wav2vec2-large-en-de''': ( '''https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json''' ), }, '''tokenizer_config_file''': { '''facebook/s2t-wav2vec2-large-en-de''': ( '''https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json''' ), }, '''merges_file''': { '''facebook/s2t-wav2vec2-large-en-de''': ( '''https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt''' ), }, } _UpperCAmelCase : str = '''</w>''' _UpperCAmelCase : Dict = '''@@ ''' def UpperCamelCase ( lowercase_ : Union[str, Any] ) -> Dict: '''simple docstring''' lowercase =set() lowercase =word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase =char return pairs # Speech2Text2 has no max input length _UpperCAmelCase : str = {'''facebook/s2t-wav2vec2-large-en-de''': 10_24} class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<s>" , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_=False , snake_case_=None , **snake_case_ , ): super().__init__( unk_token=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , pad_token=snake_case_ , do_lower_case=snake_case_ , **snake_case_ , ) lowercase =do_lower_case with open(snake_case_ , encoding='''utf-8''' ) as vocab_handle: lowercase =json.load(snake_case_ ) lowercase ={v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f'No merges files provided. {self.__class__.__name__} can only be used for decoding.' ) lowercase =None lowercase =None else: with open(snake_case_ , encoding='''utf-8''' ) as merges_handle: lowercase =merges_handle.read().split('''\n''' )[:-1] lowercase =[tuple(merge.split()[:2] ) for merge in merges] lowercase =dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) lowercase ={} @property def _A( self ): return len(self.decoder ) def _A( self ): return dict(self.encoder , **self.added_tokens_encoder ) def _A( self , snake_case_ ): lowercase =tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] lowercase =get_pairs(snake_case_ ) if not pairs: return token while True: lowercase =min(snake_case_ , key=lambda snake_case_ : self.bpe_ranks.get(snake_case_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowercase , lowercase =bigram lowercase =[] lowercase =0 while i < len(snake_case_ ): try: lowercase =word.index(snake_case_ , snake_case_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase =j if word[i] == first and i < len(snake_case_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase =tuple(snake_case_ ) lowercase =new_word if len(snake_case_ ) == 1: break else: lowercase =get_pairs(snake_case_ ) lowercase =''' '''.join(snake_case_ ) if word == "\n " + BPE_TOKEN_MERGES: lowercase ='''\n''' + BPE_TOKEN_MERGES if word.endswith(snake_case_ ): lowercase =word.replace(snake_case_ , '''''' ) lowercase =word.replace(''' ''' , snake_case_ ) lowercase =word return word def _A( self , snake_case_ ): if self.bpe_ranks is None: raise ValueError( '''This tokenizer was instantiated without a `merges.txt` file, so''' ''' that it can only be used for decoding, not for encoding.''' '''Make sure to provide `merges.txt` file at instantiation to enable ''' '''encoding.''' ) if self.do_lower_case: lowercase =text.lower() lowercase =text.split() lowercase =[] for token in text: if token: split_tokens.extend(list(self.bpe(snake_case_ ).split(''' ''' ) ) ) return split_tokens def _A( self , snake_case_ ): return self.encoder.get(snake_case_ , self.encoder.get(self.unk_token ) ) def _A( self , snake_case_ ): lowercase =self.decoder.get(snake_case_ , self.unk_token ) return result def _A( self , snake_case_ ): lowercase =''' '''.join(snake_case_ ) # make sure @@ tokens are concatenated lowercase =''''''.join(string.split(snake_case_ ) ) return string def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(snake_case_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case_ , ensure_ascii=snake_case_ ) + '''\n''' ) lowercase =0 if self.bpe_ranks is None: return (vocab_file,) with open(snake_case_ , '''w''' , encoding='''utf-8''' ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case_ : kv[1] ): if index != token_index: logger.warning( f'Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.' ''' Please check that the tokenizer is not corrupted!''' ) lowercase =token_index writer.write(''' '''.join(snake_case_ ) + '''\n''' ) index += 1 return (vocab_file, merges_file)
72
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
1
'''simple docstring''' from ..utils import DummyObject, requires_backends class __magic_name__ ( metaclass=__SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['note_seq'] def __init__( self , *snake_case_ , **snake_case_ ): requires_backends(self , ['''note_seq'''] ) @classmethod def _A( cls , *snake_case_ , **snake_case_ ): requires_backends(cls , ['''note_seq'''] ) @classmethod def _A( cls , *snake_case_ , **snake_case_ ): requires_backends(cls , ['''note_seq'''] )
72
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __magic_name__ ( unittest.TestCase ): def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=18 , snake_case_=30 , snake_case_=4_00 , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=[0.5, 0.5, 0.5] , snake_case_=[0.5, 0.5, 0.5] , ): lowercase =size if size is not None else {'''shortest_edge''': 18} lowercase =crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} lowercase =parent lowercase =batch_size lowercase =num_channels lowercase =image_size lowercase =min_resolution lowercase =max_resolution lowercase =do_resize lowercase =size lowercase =do_center_crop lowercase =crop_size lowercase =do_normalize lowercase =image_mean lowercase =image_std def _A( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = LevitImageProcessor if is_vision_available() else None def _A( self ): lowercase =LevitImageProcessingTester(self ) @property def _A( self ): return self.image_processor_tester.prepare_image_processor_dict() def _A( self ): lowercase =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , '''image_mean''' ) ) self.assertTrue(hasattr(snake_case_ , '''image_std''' ) ) self.assertTrue(hasattr(snake_case_ , '''do_normalize''' ) ) self.assertTrue(hasattr(snake_case_ , '''do_resize''' ) ) self.assertTrue(hasattr(snake_case_ , '''do_center_crop''' ) ) self.assertTrue(hasattr(snake_case_ , '''size''' ) ) def _A( self ): lowercase =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) lowercase =self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def _A( self ): pass def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _A( self ): # Initialize image_processing lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input lowercase =image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase =image_processing(snake_case_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
72
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'vit_msn' def __init__( self , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_=30_72 , snake_case_="gelu" , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=1E-06 , snake_case_=2_24 , snake_case_=16 , snake_case_=3 , snake_case_=True , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =initializer_range lowercase =layer_norm_eps lowercase =image_size lowercase =patch_size lowercase =num_channels lowercase =qkv_bias
72
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
1
'''simple docstring''' # This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> List[Any]: '''simple docstring''' lowercase =multiprocessing.Manager() lowercase =manager.list() lowercase =multiprocessing.Process(target=lowercase_ , args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append('''timed out''' ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def UpperCamelCase ( lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : List[str] ) -> List[str]: '''simple docstring''' with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil lowercase =shutil.rmtree lowercase =os.rmdir lowercase =os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: lowercase ={} with swallow_io(): with time_limit(lowercase_ ): exec(lowercase_ , lowercase_ ) result.append('''passed''' ) except TimeoutException: result.append('''timed out''' ) except BaseException as e: result.append(f'failed: {e}' ) # Needed for cleaning up. lowercase =rmtree lowercase =rmdir lowercase =chdir @contextlib.contextmanager def UpperCamelCase ( lowercase_ : List[str] ) -> Optional[int]: '''simple docstring''' def signal_handler(lowercase_ : Optional[int] , lowercase_ : str ): raise TimeoutException('''Timed out!''' ) signal.setitimer(signal.ITIMER_REAL , lowercase_ ) signal.signal(signal.SIGALRM , lowercase_ ) try: yield finally: signal.setitimer(signal.ITIMER_REAL , 0 ) @contextlib.contextmanager def UpperCamelCase ( ) -> Dict: '''simple docstring''' lowercase =WriteOnlyStringIO() with contextlib.redirect_stdout(lowercase_ ): with contextlib.redirect_stderr(lowercase_ ): with redirect_stdin(lowercase_ ): yield @contextlib.contextmanager def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as dirname: with chdir(lowercase_ ): yield dirname class __magic_name__ ( __SCREAMING_SNAKE_CASE ): pass class __magic_name__ ( io.StringIO ): def _A( self , *snake_case_ , **snake_case_ ): raise OSError def _A( self , *snake_case_ , **snake_case_ ): raise OSError def _A( self , *snake_case_ , **snake_case_ ): raise OSError def _A( self , *snake_case_ , **snake_case_ ): return False class __magic_name__ ( contextlib._RedirectStream ): # type: ignore UpperCamelCase__ = 'stdin' @contextlib.contextmanager def UpperCamelCase ( lowercase_ : List[str] ) -> Optional[Any]: '''simple docstring''' if root == ".": yield return lowercase =os.getcwd() os.chdir(lowercase_ ) try: yield except BaseException as exc: raise exc finally: os.chdir(lowercase_ ) def UpperCamelCase ( lowercase_ : str=None ) -> Tuple: '''simple docstring''' if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS , (maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA , (maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK , (maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins lowercase =None lowercase =None import os lowercase ='''1''' lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None import shutil lowercase =None lowercase =None lowercase =None import subprocess lowercase =None # type: ignore lowercase =None import sys lowercase =None lowercase =None lowercase =None lowercase =None lowercase =None
72
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
1
'''simple docstring''' from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('''socket.socket''' ) @patch('''builtins.open''' ) def UpperCamelCase ( lowercase_ : int , lowercase_ : List[str] ) -> Dict: '''simple docstring''' lowercase =Mock() lowercase =conn, Mock() lowercase =iter([1, None] ) lowercase =lambda lowercase_ : next(lowercase_ ) # ===== invoke ===== send_file(filename='''mytext.txt''' , testing=lowercase_ ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
72
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
1
'''simple docstring''' import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def UpperCamelCase ( lowercase_ : Union[str, Any] ) -> Any: '''simple docstring''' lowercase =[ '''decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase , lowercase =emb.weight.shape lowercase =nn.Linear(lowercase_ , lowercase_ , bias=lowercase_ ) lowercase =emb.weight.data return lin_layer def UpperCamelCase ( lowercase_ : str ) -> int: '''simple docstring''' lowercase =torch.load(lowercase_ , map_location='''cpu''' ) lowercase =Namespace(**checkpoint['''cfg''']['''model'''] ) lowercase =checkpoint['''model'''] remove_ignore_keys_(lowercase_ ) lowercase =state_dict['''decoder.embed_tokens.weight'''].shape[0] lowercase ={key.replace('''decoder''' , '''model''' ): val for key, val in state_dict.items()} lowercase =XGLMConfig( vocab_size=lowercase_ , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''gelu''' , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) lowercase =XGLMForCausalLM(lowercase_ ) lowercase =model.load_state_dict(lowercase_ , strict=lowercase_ ) print(lowercase_ ) lowercase =make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument('''fairseq_path''', type=str, help='''path to a model.pt on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _UpperCAmelCase : Dict = parser.parse_args() _UpperCAmelCase : Optional[int] = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
72
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
1
'''simple docstring''' import math class __magic_name__ : def _A( self , snake_case_ , snake_case_ ): lowercase =0.0 lowercase =0.0 for i in range(len(snake_case_ ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): for i in range(len(snake_case_ ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def UpperCamelCase ( ) -> None: '''simple docstring''' lowercase =[[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) lowercase =[[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training lowercase =SelfOrganizingMap() lowercase =3 lowercase =0.5 for _ in range(lowercase_ ): for j in range(len(lowercase_ ) ): # training sample lowercase =training_samples[j] # Compute the winning vector lowercase =self_organizing_map.get_winner(lowercase_ , lowercase_ ) # Update the winning vector lowercase =self_organizing_map.update(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # classify test sample lowercase =[0, 0, 0, 1] lowercase =self_organizing_map.get_winner(lowercase_ , lowercase_ ) # results print(f'Clusters that the test sample belongs to : {winner}' ) print(f'Weights that have been trained : {weights}' ) # running the main() function if __name__ == "__main__": main()
72
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __magic_name__ : def __init__( self , snake_case_ , snake_case_=12 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=0.02 , snake_case_=0 , snake_case_=None , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =is_training lowercase =use_input_mask lowercase =use_labels lowercase =vocab_size lowercase =hidden_size lowercase =projection_dim lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =dropout lowercase =attention_dropout lowercase =max_position_embeddings lowercase =initializer_range lowercase =scope lowercase =bos_token_id def _A( self ): lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase =None if self.use_input_mask: lowercase =random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase =input_mask.numpy() lowercase , lowercase =input_mask.shape lowercase =np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(snake_case_ ): lowercase =1 lowercase =0 lowercase =self.get_config() return config, input_ids, tf.convert_to_tensor(snake_case_ ) def _A( self ): return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =TFBlipTextModel(config=snake_case_ ) lowercase =model(snake_case_ , attention_mask=snake_case_ , training=snake_case_ ) lowercase =model(snake_case_ , training=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _A( self ): lowercase =self.prepare_config_and_inputs() lowercase , lowercase , lowercase =config_and_inputs lowercase ={'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =BlipTextModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _A( self ): pass def _A( self ): pass @unittest.skip(reason='''Blip does not use inputs_embeds''' ) def _A( self ): pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def _A( self ): pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def _A( self ): pass @slow def _A( self ): for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =TFBlipTextModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self , snake_case_=True ): super().test_pt_tf_model_equivalence(allow_missing_keys=snake_case_ )
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
1
'''simple docstring''' import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = DownBlockaD # noqa F405 UpperCamelCase__ = 'down' def _A( self ): lowercase =[-0.02_32, -0.98_69, 0.80_54, -0.06_37, -0.16_88, -1.42_64, 0.44_70, -1.33_94, 0.09_04] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = ResnetDownsampleBlockaD # noqa F405 UpperCamelCase__ = 'down' def _A( self ): lowercase =[0.07_10, 0.24_10, -0.73_20, -1.07_57, -1.13_43, 0.35_40, -0.01_33, -0.25_76, 0.09_48] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnDownBlockaD # noqa F405 UpperCamelCase__ = 'down' def _A( self ): lowercase =[0.06_36, 0.89_64, -0.62_34, -1.01_31, 0.08_44, 0.49_35, 0.34_37, 0.09_11, -0.29_57] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = CrossAttnDownBlockaD # noqa F405 UpperCamelCase__ = 'down' def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict def _A( self ): lowercase =[0.22_38, -0.73_96, -0.22_55, -0.38_29, 0.19_25, 1.16_65, 0.06_03, -0.72_95, 0.19_83] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = SimpleCrossAttnDownBlockaD # noqa F405 UpperCamelCase__ = 'down' @property def _A( self ): return super().get_dummy_input(include_encoder_hidden_states=snake_case_ ) def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict @unittest.skipIf(torch_device == '''mps''' , '''MPS result is not consistent''' ) def _A( self ): lowercase =[0.79_21, -0.09_92, -0.19_62, -0.76_95, -0.42_42, 0.78_04, 0.47_37, 0.27_65, 0.33_38] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = SkipDownBlockaD # noqa F405 UpperCamelCase__ = 'down' @property def _A( self ): return super().get_dummy_input(include_skip_sample=snake_case_ ) def _A( self ): lowercase =[-0.08_45, -0.20_87, -0.24_65, 0.09_71, 0.19_00, -0.04_84, 0.26_64, 0.41_79, 0.50_69] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnSkipDownBlockaD # noqa F405 UpperCamelCase__ = 'down' @property def _A( self ): return super().get_dummy_input(include_skip_sample=snake_case_ ) def _A( self ): lowercase =[0.55_39, 0.16_09, 0.49_24, 0.05_37, -0.19_95, 0.40_50, 0.09_79, -0.27_21, -0.06_42] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = DownEncoderBlockaD # noqa F405 UpperCamelCase__ = 'down' @property def _A( self ): return super().get_dummy_input(include_temb=snake_case_ ) def _A( self ): lowercase ={ '''in_channels''': 32, '''out_channels''': 32, } lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase =[1.11_02, 0.53_02, 0.48_72, -0.00_23, -0.80_42, 0.04_83, -0.34_89, -0.56_32, 0.76_26] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnDownEncoderBlockaD # noqa F405 UpperCamelCase__ = 'down' @property def _A( self ): return super().get_dummy_input(include_temb=snake_case_ ) def _A( self ): lowercase ={ '''in_channels''': 32, '''out_channels''': 32, } lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase =[0.89_66, -0.14_86, 0.85_68, 0.81_41, -0.90_46, -0.13_42, -0.09_72, -0.74_17, 0.15_38] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetMidBlockaD # noqa F405 UpperCamelCase__ = 'mid' def _A( self ): lowercase ={ '''in_channels''': 32, '''temb_channels''': 1_28, } lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase =[-0.10_62, 1.72_48, 0.34_94, 1.45_69, -0.09_10, -1.24_21, -0.99_84, 0.67_36, 1.00_28] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetMidBlockaDCrossAttn # noqa F405 UpperCamelCase__ = 'mid' def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict def _A( self ): lowercase =[0.01_87, 2.42_20, 0.44_84, 1.12_03, -0.61_21, -1.51_22, -0.82_70, 0.78_51, 1.83_35] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UNetMidBlockaDSimpleCrossAttn # noqa F405 UpperCamelCase__ = 'mid' @property def _A( self ): return super().get_dummy_input(include_encoder_hidden_states=snake_case_ ) def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict def _A( self ): lowercase =[0.71_43, 1.99_74, 0.54_48, 1.39_77, 0.12_82, -1.12_37, -1.42_38, 0.55_30, 0.88_80] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) def _A( self ): lowercase =[-0.20_41, -0.41_65, -0.30_22, 0.00_41, -0.66_28, -0.70_53, 0.19_28, -0.03_25, 0.05_23] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = ResnetUpsampleBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) def _A( self ): lowercase =[0.22_87, 0.35_49, -0.13_46, 0.47_97, -0.17_15, -0.96_49, 0.73_05, -0.58_64, -0.62_44] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = CrossAttnUpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict def _A( self ): lowercase =[-0.14_03, -0.35_15, -0.04_20, -0.14_25, 0.31_67, 0.50_94, -0.21_81, 0.59_31, 0.55_82] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = SimpleCrossAttnUpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ , include_encoder_hidden_states=snake_case_ ) def _A( self ): lowercase , lowercase =super().prepare_init_args_and_inputs_for_common() lowercase =32 return init_dict, inputs_dict def _A( self ): lowercase =[0.26_45, 0.14_80, 0.09_09, 0.80_44, -0.97_58, -0.90_83, 0.09_94, -1.14_53, -0.74_02] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnUpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) @unittest.skipIf(torch_device == '''mps''' , '''MPS result is not consistent''' ) def _A( self ): lowercase =[0.09_79, 0.13_26, 0.00_21, 0.06_59, 0.22_49, 0.00_59, 0.11_32, 0.59_52, 0.10_33] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = SkipUpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) def _A( self ): lowercase =[-0.08_93, -0.12_34, -0.15_06, -0.03_32, 0.01_23, -0.02_11, 0.05_66, 0.01_43, 0.03_62] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnSkipUpBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_res_hidden_states_tuple=snake_case_ ) def _A( self ): lowercase =[0.03_61, 0.06_17, 0.27_87, -0.03_50, 0.03_42, 0.34_21, -0.08_43, 0.09_13, 0.30_15] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = UpDecoderBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_temb=snake_case_ ) def _A( self ): lowercase ={'''in_channels''': 32, '''out_channels''': 32} lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase =[0.44_04, 0.19_98, -0.98_86, -0.33_20, -0.31_28, -0.70_34, -0.69_55, -0.23_38, -0.31_37] super().test_output(snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = AttnUpDecoderBlockaD # noqa F405 UpperCamelCase__ = 'up' @property def _A( self ): return super().get_dummy_input(include_temb=snake_case_ ) def _A( self ): lowercase ={'''in_channels''': 32, '''out_channels''': 32} lowercase =self.dummy_input return init_dict, inputs_dict def _A( self ): lowercase =[0.67_38, 0.44_91, 0.10_55, 1.07_10, 0.73_16, 0.33_39, 0.33_52, 0.10_23, 0.35_68] super().test_output(snake_case_ )
72
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
1
'''simple docstring''' from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
72
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) _UpperCAmelCase : List[str] = logging.getLogger(__name__) def UpperCamelCase ( ) -> str: '''simple docstring''' lowercase =argparse.ArgumentParser( description='''Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).''' ) parser.add_argument('''--file_path''' , type=lowercase_ , default='''data/dump.txt''' , help='''The path to the data.''' ) parser.add_argument('''--tokenizer_type''' , type=lowercase_ , default='''bert''' , choices=['''bert''', '''roberta''', '''gpt2'''] ) parser.add_argument('''--tokenizer_name''' , type=lowercase_ , default='''bert-base-uncased''' , help='''The tokenizer to use.''' ) parser.add_argument('''--dump_file''' , type=lowercase_ , default='''data/dump''' , help='''The dump file prefix.''' ) lowercase =parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": lowercase =BertTokenizer.from_pretrained(args.tokenizer_name ) lowercase =tokenizer.special_tokens_map['''cls_token'''] # `[CLS]` lowercase =tokenizer.special_tokens_map['''sep_token'''] # `[SEP]` elif args.tokenizer_type == "roberta": lowercase =RobertaTokenizer.from_pretrained(args.tokenizer_name ) lowercase =tokenizer.special_tokens_map['''cls_token'''] # `<s>` lowercase =tokenizer.special_tokens_map['''sep_token'''] # `</s>` elif args.tokenizer_type == "gpt2": lowercase =GPTaTokenizer.from_pretrained(args.tokenizer_name ) lowercase =tokenizer.special_tokens_map['''bos_token'''] # `<|endoftext|>` lowercase =tokenizer.special_tokens_map['''eos_token'''] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , '''r''' , encoding='''utf8''' ) as fp: lowercase =fp.readlines() logger.info('''Start encoding''' ) logger.info(f'{len(lowercase_ )} examples to process.' ) lowercase =[] lowercase =0 lowercase =1_0_0_0_0 lowercase =time.time() for text in data: lowercase =f'{bos} {text.strip()} {sep}' lowercase =tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) rslt.append(lowercase_ ) iter += 1 if iter % interval == 0: lowercase =time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) lowercase =time.time() logger.info('''Finished binarization''' ) logger.info(f'{len(lowercase_ )} examples processed.' ) lowercase =f'{args.dump_file}.{args.tokenizer_name}.pickle' lowercase =tokenizer.vocab_size if vocab_size < (1 << 1_6): lowercase =[np.uintaa(lowercase_ ) for d in rslt] else: lowercase =[np.intaa(lowercase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(lowercase_ , '''wb''' ) as handle: pickle.dump(rslt_ , lowercase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
72
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
1
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Any = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : Optional[int] ) -> Any: '''simple docstring''' inspect_dataset(lowercase_ , lowercase_ ) lowercase =path + '''.py''' assert script_name in os.listdir(lowercase_ ) assert "__pycache__" not in os.listdir(lowercase_ ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : Any ) -> Any: '''simple docstring''' inspect_metric(lowercase_ , lowercase_ ) lowercase =path + '''.py''' assert script_name in os.listdir(lowercase_ ) assert "__pycache__" not in os.listdir(lowercase_ ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def UpperCamelCase ( lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : Optional[int] ) -> List[str]: '''simple docstring''' lowercase =get_dataset_config_info(lowercase_ , config_name=lowercase_ ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def UpperCamelCase ( lowercase_ : int , lowercase_ : List[str] , lowercase_ : Optional[int] ) -> Tuple: '''simple docstring''' with pytest.raises(lowercase_ ): get_dataset_config_info(lowercase_ , config_name=lowercase_ ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> int: '''simple docstring''' lowercase =get_dataset_config_names(lowercase_ ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def UpperCamelCase ( lowercase_ : Any , lowercase_ : str , lowercase_ : Tuple ) -> Tuple: '''simple docstring''' lowercase =get_dataset_infos(lowercase_ ) assert list(infos.keys() ) == expected_configs lowercase =expected_configs[0] assert expected_config in infos lowercase =infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Tuple , lowercase_ : List[str] ) -> List[str]: '''simple docstring''' lowercase =get_dataset_infos(lowercase_ ) assert expected_config in infos lowercase =infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : Any , lowercase_ : int ) -> List[str]: '''simple docstring''' with pytest.raises(lowercase_ ): get_dataset_split_names(lowercase_ , config_name=lowercase_ )
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
1
'''simple docstring''' def UpperCamelCase ( lowercase_ : list , lowercase_ : list , lowercase_ : int ) -> int: '''simple docstring''' if len(lowercase_ ) != len(lowercase_ ): raise ValueError('''The length of profit and weight must be same.''' ) if max_weight <= 0: raise ValueError('''max_weight must greater than zero.''' ) if any(p < 0 for p in profit ): raise ValueError('''Profit can not be negative.''' ) if any(w < 0 for w in weight ): raise ValueError('''Weight can not be negative.''' ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. lowercase =[p / w for p, w in zip(lowercase_ , lowercase_ )] # Creating a copy of the list and sorting profit/weight in ascending order lowercase =sorted(lowercase_ ) # declaring useful variables lowercase =len(lowercase_ ) lowercase =0 lowercase =0 lowercase =0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight lowercase =sorted_profit_by_weight[length - i - 1] lowercase =profit_by_weight.index(lowercase_ ) lowercase =-1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( '''Input profits, weights, and then max_weight (all positive ints) separated by ''' '''spaces.''' ) _UpperCAmelCase : List[str] = [int(x) for x in input('''Input profits separated by spaces: ''').split()] _UpperCAmelCase : Optional[int] = [int(x) for x in input('''Input weights separated by spaces: ''').split()] _UpperCAmelCase : Union[str, Any] = int(input('''Max weight allowed: ''')) # Function Call calc_profit(profit, weight, max_weight)
72
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
1
'''simple docstring''' import doctest from collections import deque import numpy as np class __magic_name__ : def __init__( self ): lowercase =[2, 1, 2, -1] lowercase =[1, 2, 3, 4] def _A( self ): lowercase =len(self.first_signal ) lowercase =len(self.second_signal ) lowercase =max(snake_case_ , snake_case_ ) # create a zero matrix of max_length x max_length lowercase =[[0] * max_length for i in range(snake_case_ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(snake_case_ ): lowercase =deque(self.second_signal ) rotated_signal.rotate(snake_case_ ) for j, item in enumerate(snake_case_ ): matrix[i][j] += item # multiply the matrix with the first signal lowercase =np.matmul(np.transpose(snake_case_ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(snake_case_ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
72
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
1
'''simple docstring''' import argparse import os import re _UpperCAmelCase : Optional[Any] = '''src/transformers/models/auto''' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict _UpperCAmelCase : List[str] = re.compile(r'''[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict''') # re pattern that matches identifiers in mappings _UpperCAmelCase : Optional[int] = re.compile(r'''\s*\(\s*"(\S[^"]+)"''') def UpperCamelCase ( lowercase_ : Dict , lowercase_ : bool = False ) -> List[Any]: '''simple docstring''' with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: lowercase =f.read() lowercase =content.split('''\n''' ) lowercase =[] lowercase =0 while line_idx < len(lowercase_ ): if _re_intro_mapping.search(lines[line_idx] ) is not None: lowercase =len(re.search(R'''^(\s*)\S''' , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(''' ''' * indent + '''(''' ): new_lines.append(lines[line_idx] ) line_idx += 1 lowercase =[] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": lowercase =line_idx while not lines[line_idx].startswith(''' ''' * indent + ''')''' ): line_idx += 1 blocks.append('''\n'''.join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers lowercase =sorted(lowercase_ , key=lambda lowercase_ : _re_identifier.search(lowercase_ ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write('''\n'''.join(lowercase_ ) ) elif "\n".join(lowercase_ ) != content: return True def UpperCamelCase ( lowercase_ : bool = False ) -> str: '''simple docstring''' lowercase =[os.path.join(lowercase_ , lowercase_ ) for f in os.listdir(lowercase_ ) if f.endswith('''.py''' )] lowercase =[sort_auto_mapping(lowercase_ , overwrite=lowercase_ ) for fname in fnames] if not overwrite and any(lowercase_ ): lowercase =[f for f, d in zip(lowercase_ , lowercase_ ) if d] raise ValueError( f'The following files have auto mappings that need sorting: {", ".join(lowercase_ )}. Run `make style` to fix' ''' this.''' ) if __name__ == "__main__": _UpperCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') _UpperCAmelCase : int = parser.parse_args() sort_all_auto_mappings(not args.check_only)
72
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' from collections import Counter from timeit import timeit def UpperCamelCase ( lowercase_ : str = "" , ) -> bool: '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(''' ''' , '''''' ).lower() ).values() ) < 2 def UpperCamelCase ( lowercase_ : str = "" ) -> bool: '''simple docstring''' if len(lowercase_ ) == 0: return True lowercase =input_str.replace(''' ''' , '''''' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowercase ={} for character in lower_case_input_str: lowercase =character_freq_dict.get(lowercase_ , 0 ) + 1 lowercase =0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def UpperCamelCase ( lowercase_ : str = "" ) -> None: '''simple docstring''' print('''\nFor string = ''' , lowercase_ , ''':''' ) print( '''> can_string_be_rearranged_as_palindrome_counter()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome_counter(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome_counter(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) print( '''> can_string_be_rearranged_as_palindrome()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) if __name__ == "__main__": _UpperCAmelCase : List[str] = input( '''Enter string to determine if it can be rearranged as a palindrome or not: ''' ).strip() benchmark(check_str) _UpperCAmelCase : str = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
72
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCAmelCase : Optional[int] = { '''configuration_blenderbot''': [ '''BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotConfig''', '''BlenderbotOnnxConfig''', ], '''tokenization_blenderbot''': ['''BlenderbotTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : str = ['''BlenderbotTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ '''BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotForCausalLM''', '''BlenderbotForConditionalGeneration''', '''BlenderbotModel''', '''BlenderbotPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ '''TFBlenderbotForConditionalGeneration''', '''TFBlenderbotModel''', '''TFBlenderbotPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ '''FlaxBlenderbotForConditionalGeneration''', '''FlaxBlenderbotModel''', '''FlaxBlenderbotPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys _UpperCAmelCase : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
1
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
1
'''simple docstring''' from __future__ import annotations import typing from collections import Counter def UpperCamelCase ( lowercase_ : int ) -> typing.Counter[int]: '''simple docstring''' lowercase =Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(lowercase_ , max_perimeter + 1 ): lowercase =(base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(lowercase_ ): lowercase =int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' lowercase =pythagorean_triple(lowercase_ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(F"""Perimeter {solution()} has maximum solutions""")
72
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
1
'''simple docstring''' from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _UpperCAmelCase : List[Any] = logging.get_logger(__name__) @add_end_docstrings(__SCREAMING_SNAKE_CASE ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): super().__init__(*snake_case_ , **snake_case_ ) self.check_model_type(snake_case_ ) def _A( self , snake_case_=None , snake_case_=None , snake_case_=None , **snake_case_ ): lowercase , lowercase ={}, {} if padding is not None: lowercase =padding if truncation is not None: lowercase =truncation if top_k is not None: lowercase =top_k return preprocess_params, {}, postprocess_params def __call__( self , snake_case_ , snake_case_ = None , **snake_case_ ): if isinstance(snake_case_ , (Image.Image, str) ) and isinstance(snake_case_ , snake_case_ ): lowercase ={'''image''': image, '''question''': question} else: lowercase =image lowercase =super().__call__(snake_case_ , **snake_case_ ) return results def _A( self , snake_case_ , snake_case_=False , snake_case_=False ): lowercase =load_image(inputs['''image'''] ) lowercase =self.tokenizer( inputs['''question'''] , return_tensors=self.framework , padding=snake_case_ , truncation=snake_case_ ) lowercase =self.image_processor(images=snake_case_ , return_tensors=self.framework ) model_inputs.update(snake_case_ ) return model_inputs def _A( self , snake_case_ ): lowercase =self.model(**snake_case_ ) return model_outputs def _A( self , snake_case_ , snake_case_=5 ): if top_k > self.model.config.num_labels: lowercase =self.model.config.num_labels if self.framework == "pt": lowercase =model_outputs.logits.sigmoid()[0] lowercase , lowercase =probs.topk(snake_case_ ) else: raise ValueError(f'Unsupported framework: {self.framework}' ) lowercase =scores.tolist() lowercase =ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(snake_case_ , snake_case_ )]
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' import re def UpperCamelCase ( lowercase_ : str ) -> list: '''simple docstring''' return [char.split() for char in re.split(R'''[^ a-z A-Z 0-9 \s]''' , str_ )] def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' lowercase =split_input(str_ ) return "".join( [''''''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def UpperCamelCase ( lowercase_ : str , lowercase_ : bool , lowercase_ : str ) -> str: '''simple docstring''' try: lowercase =split_input(lowercase_ ) if upper: lowercase =''''''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: lowercase =''''''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' return to_simple_case(lowercase_ ) def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' try: lowercase =to_simple_case(lowercase_ ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def UpperCamelCase ( lowercase_ : str , lowercase_ : bool ) -> str: '''simple docstring''' return to_complex_case(lowercase_ , lowercase_ , '''_''' ) def UpperCamelCase ( lowercase_ : str , lowercase_ : bool ) -> str: '''simple docstring''' return to_complex_case(lowercase_ , lowercase_ , '''-''' ) if __name__ == "__main__": __import__('''doctest''').testmod()
72
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' from sklearn.metrics import matthews_corrcoef import datasets _UpperCAmelCase : Any = ''' Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] ''' _UpperCAmelCase : str = ''' Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results[\'matthews_correlation\'], 2)) -0.25 ''' _UpperCAmelCase : Union[str, Any] = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=[ '''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html''' ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None ): return { "matthews_correlation": float(matthews_corrcoef(snake_case_ , snake_case_ , sample_weight=snake_case_ ) ), }
72
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
1
'''simple docstring''' from string import ascii_uppercase _UpperCAmelCase : List[str] = {str(ord(c) - 55): c for c in ascii_uppercase} def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): raise TypeError('''int() can\'t convert non-string with explicit base''' ) if num < 0: raise ValueError('''parameter must be positive int''' ) if isinstance(lowercase_ , lowercase_ ): raise TypeError('''\'str\' object cannot be interpreted as an integer''' ) if isinstance(lowercase_ , lowercase_ ): raise TypeError('''\'float\' object cannot be interpreted as an integer''' ) if base in (0, 1): raise ValueError('''base must be >= 2''' ) if base > 3_6: raise ValueError('''base must be <= 36''' ) lowercase ='''''' lowercase =0 lowercase =0 while div != 1: lowercase , lowercase =divmod(lowercase_ , lowercase_ ) if base >= 1_1 and 9 < mod < 3_6: lowercase =ALPHABET_VALUES[str(lowercase_ )] else: lowercase =str(lowercase_ ) new_value += actual_value lowercase =num // base lowercase =div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(lowercase_ ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(10_00): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
72
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
1
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCAmelCase : int = { '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'gpt_neo' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , snake_case_=5_02_57 , snake_case_=20_48 , snake_case_=20_48 , snake_case_=24 , snake_case_=[[["global", "local"], 12]] , snake_case_=16 , snake_case_=None , snake_case_=2_56 , snake_case_="gelu_new" , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_=True , snake_case_=5_02_56 , snake_case_=5_02_56 , **snake_case_ , ): lowercase =vocab_size lowercase =max_position_embeddings lowercase =hidden_size lowercase =num_layers lowercase =num_heads lowercase =intermediate_size lowercase =window_size lowercase =activation_function lowercase =resid_dropout lowercase =embed_dropout lowercase =attention_dropout lowercase =classifier_dropout lowercase =layer_norm_epsilon lowercase =initializer_range lowercase =use_cache lowercase =bos_token_id lowercase =eos_token_id lowercase =attention_types lowercase =self.expand_attention_types_params(snake_case_ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' f'but is `len(config.attention_layers) = {len(self.attention_layers )}`, ' f'`config.num_layers = {self.num_layers}`. ' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''' ) super().__init__(bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) @staticmethod def _A( snake_case_ ): lowercase =[] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : str , lowercase_ : str , lowercase_ : Optional[int] ) -> str: '''simple docstring''' import torch lowercase =input.size() lowercase =len(lowercase_ ) lowercase =shape[dimension] lowercase =torch.arange(0 , lowercase_ , lowercase_ ) lowercase =torch.div(sizedim - size , lowercase_ , rounding_mode='''floor''' ) + 1 lowercase =torch.arange(lowercase_ ) + low_indices[:min_length][:, None] lowercase =[slice(lowercase_ )] * rank lowercase =indices lowercase =input[s] lowercase =list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowercase_ ) def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Union[str, Any] ) -> List[str]: '''simple docstring''' import torch lowercase =torch.arange(1 , lowercase_ ) lowercase =torch.remainder(lowercase_ , lowercase_ ) lowercase =remainders == 0 lowercase =candidates[divisor_indices] lowercase =torch.max(lowercase_ ) return largest_divisor, torch.div(lowercase_ , lowercase_ , rounding_mode='''floor''' ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property def _A( self ): lowercase =OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) lowercase ={0: '''batch''', 1: '''past_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''sequence'''} return common_inputs @property def _A( self ): return self._config.num_heads def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =super(snake_case_ , self ).generate_dummy_inputs( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) # We need to order the input in the way they appears in the forward() lowercase =OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase =( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(self.num_layers ) ] lowercase =common_inputs['''attention_mask'''] if self.use_past: lowercase =ordered_inputs['''attention_mask'''].dtype lowercase =torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) return ordered_inputs @property def _A( self ): return 13
72
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
1
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCAmelCase : str = logging.get_logger(__name__) @add_end_docstrings(__SCREAMING_SNAKE_CASE ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , **snake_case_ ): super().__init__(**snake_case_ ) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , '''vision''' ) self.check_model_type(snake_case_ ) def __call__( self , snake_case_ , snake_case_ = None , **snake_case_ , ): if "text_queries" in kwargs: lowercase =kwargs.pop('''text_queries''' ) if isinstance(snake_case_ , (str, Image.Image) ): lowercase ={'''image''': image, '''candidate_labels''': candidate_labels} else: lowercase =image lowercase =super().__call__(snake_case_ , **snake_case_ ) return results def _A( self , **snake_case_ ): lowercase ={} if "threshold" in kwargs: lowercase =kwargs['''threshold'''] if "top_k" in kwargs: lowercase =kwargs['''top_k'''] return {}, {}, postprocess_params def _A( self , snake_case_ ): lowercase =load_image(inputs['''image'''] ) lowercase =inputs['''candidate_labels'''] if isinstance(snake_case_ , snake_case_ ): lowercase =candidate_labels.split(''',''' ) lowercase =torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(snake_case_ ): lowercase =self.tokenizer(snake_case_ , return_tensors=self.framework ) lowercase =self.image_processor(snake_case_ , return_tensors=self.framework ) yield { "is_last": i == len(snake_case_ ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _A( self , snake_case_ ): lowercase =model_inputs.pop('''target_size''' ) lowercase =model_inputs.pop('''candidate_label''' ) lowercase =model_inputs.pop('''is_last''' ) lowercase =self.model(**snake_case_ ) lowercase ={'''target_size''': target_size, '''candidate_label''': candidate_label, '''is_last''': is_last, **outputs} return model_outputs def _A( self , snake_case_ , snake_case_=0.1 , snake_case_=None ): lowercase =[] for model_output in model_outputs: lowercase =model_output['''candidate_label'''] lowercase =BaseModelOutput(snake_case_ ) lowercase =self.image_processor.post_process_object_detection( outputs=snake_case_ , threshold=snake_case_ , target_sizes=model_output['''target_size'''] )[0] for index in outputs["scores"].nonzero(): lowercase =outputs['''scores'''][index].item() lowercase =self._get_bounding_box(outputs['''boxes'''][index][0] ) lowercase ={'''score''': score, '''label''': label, '''box''': box} results.append(snake_case_ ) lowercase =sorted(snake_case_ , key=lambda snake_case_ : x["score"] , reverse=snake_case_ ) if top_k: lowercase =results[:top_k] return results def _A( self , snake_case_ ): if self.framework != "pt": raise ValueError('''The ZeroShotObjectDetectionPipeline is only available in PyTorch.''' ) lowercase , lowercase , lowercase , lowercase =box.int().tolist() lowercase ={ '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
72
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
1
'''simple docstring''' import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() _UpperCAmelCase : List[Any] = logging.get_logger('''transformers.models.speecht5''') _UpperCAmelCase : Optional[int] = { '''speech_encoder_prenet.layer_norm''': '''speecht5.encoder.prenet.feature_projection.layer_norm''', '''speech_encoder_prenet.post_extract_proj''': '''speecht5.encoder.prenet.feature_projection.projection''', '''speech_encoder_prenet.pos_conv.0''': '''speecht5.encoder.prenet.pos_conv_embed.conv''', '''speech_encoder_prenet.mask_emb''': '''speecht5.encoder.prenet.masked_spec_embed''', } _UpperCAmelCase : Optional[Any] = { '''text_encoder_prenet.encoder_prenet.0''': '''speecht5.encoder.prenet.embed_tokens''', '''text_encoder_prenet.encoder_prenet.1.alpha''': '''speecht5.encoder.prenet.encode_positions.alpha''', } _UpperCAmelCase : Tuple = { '''speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0''': '''speecht5.decoder.prenet.layers.0''', '''speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0''': '''speecht5.decoder.prenet.layers.1''', '''speech_decoder_prenet.decoder_prenet.0.1''': '''speecht5.decoder.prenet.final_layer''', '''speech_decoder_prenet.decoder_prenet.1.alpha''': '''speecht5.decoder.prenet.encode_positions.alpha''', '''speech_decoder_prenet.spkembs_layer.0''': '''speecht5.decoder.prenet.speaker_embeds_layer''', } _UpperCAmelCase : Optional[Any] = { '''speech_decoder_postnet.feat_out''': '''speech_decoder_postnet.feat_out''', '''speech_decoder_postnet.prob_out''': '''speech_decoder_postnet.prob_out''', '''speech_decoder_postnet.postnet.postnet.0.0''': '''speech_decoder_postnet.layers.0.conv''', '''speech_decoder_postnet.postnet.postnet.0.1''': '''speech_decoder_postnet.layers.0.batch_norm''', '''speech_decoder_postnet.postnet.postnet.1.0''': '''speech_decoder_postnet.layers.1.conv''', '''speech_decoder_postnet.postnet.postnet.1.1''': '''speech_decoder_postnet.layers.1.batch_norm''', '''speech_decoder_postnet.postnet.postnet.2.0''': '''speech_decoder_postnet.layers.2.conv''', '''speech_decoder_postnet.postnet.postnet.2.1''': '''speech_decoder_postnet.layers.2.batch_norm''', '''speech_decoder_postnet.postnet.postnet.3.0''': '''speech_decoder_postnet.layers.3.conv''', '''speech_decoder_postnet.postnet.postnet.3.1''': '''speech_decoder_postnet.layers.3.batch_norm''', '''speech_decoder_postnet.postnet.postnet.4.0''': '''speech_decoder_postnet.layers.4.conv''', '''speech_decoder_postnet.postnet.postnet.4.1''': '''speech_decoder_postnet.layers.4.batch_norm''', } _UpperCAmelCase : Tuple = { '''text_decoder_prenet.embed_tokens''': '''speecht5.decoder.prenet.embed_tokens''', } _UpperCAmelCase : Union[str, Any] = { '''text_decoder_postnet.output_projection''': '''text_decoder_postnet.lm_head''', } _UpperCAmelCase : Union[str, Any] = { '''encoder.layers.*.self_attn.k_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj''', '''encoder.layers.*.self_attn.v_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj''', '''encoder.layers.*.self_attn.q_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj''', '''encoder.layers.*.self_attn.out_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj''', '''encoder.layers.*.self_attn_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.layer_norm''', '''encoder.layers.*.fc1''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense''', '''encoder.layers.*.fc2''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense''', '''encoder.layers.*.final_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''speecht5.encoder.wrapped_encoder.layer_norm''', '''encoder.pos_emb.pe_k''': '''speecht5.encoder.wrapped_encoder.embed_positions.pe_k''', } _UpperCAmelCase : Union[str, Any] = { '''decoder.layers.*.self_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj''', '''decoder.layers.*.self_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj''', '''decoder.layers.*.self_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj''', '''decoder.layers.*.self_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj''', '''decoder.layers.*.self_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm''', '''decoder.layers.*.encoder_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj''', '''decoder.layers.*.encoder_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj''', '''decoder.layers.*.encoder_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj''', '''decoder.layers.*.encoder_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj''', '''decoder.layers.*.encoder_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm''', '''decoder.layers.*.fc1''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense''', '''decoder.layers.*.fc2''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense''', '''decoder.layers.*.final_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm''', } _UpperCAmelCase : List[Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } _UpperCAmelCase : Optional[Any] = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } _UpperCAmelCase : str = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } _UpperCAmelCase : Optional[Any] = [] _UpperCAmelCase : Union[str, Any] = [ '''encoder.version''', '''encoder.layers.*.norm_k.weight''', '''encoder.layers.*.norm_k.bias''', '''decoder.version''', '''decoder.layers.*.norm_k.weight''', '''decoder.layers.*.norm_k.bias''', '''decoder.pos_emb.pe_k''', '''speech_encoder_prenet.embed_positions._float_tensor''', '''text_decoder_prenet.embed_positions._float_tensor''', ] _UpperCAmelCase : List[str] = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''speech_decoder_prenet.*''', '''speech_decoder_postnet.*''', ] _UpperCAmelCase : int = IGNORE_KEYS + [ '''encoder.proj''', '''speech_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] _UpperCAmelCase : int = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str , lowercase_ : List[Any] , lowercase_ : List[str] ) -> Dict: '''simple docstring''' for attribute in key.split('''.''' ): lowercase =getattr(lowercase_ , lowercase_ ) if weight_type is not None: lowercase =getattr(lowercase_ , lowercase_ ).shape else: lowercase =hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": lowercase =value elif weight_type == "weight_g": lowercase =value elif weight_type == "weight_v": lowercase =value elif weight_type == "bias": lowercase =value elif weight_type == "running_mean": lowercase =value elif weight_type == "running_var": lowercase =value elif weight_type == "num_batches_tracked": lowercase =value else: lowercase =value logger.info(f'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def UpperCamelCase ( lowercase_ : str , lowercase_ : Optional[Any] ) -> Tuple: '''simple docstring''' for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: lowercase , lowercase =key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def UpperCamelCase ( lowercase_ : Any , lowercase_ : Any , lowercase_ : Optional[int] ) -> Optional[int]: '''simple docstring''' lowercase =[] if task == "s2t": lowercase =hf_model.speechta.encoder.prenet.feature_encoder lowercase =MAPPING_S2T lowercase =IGNORE_KEYS_S2T elif task == "t2s": lowercase =None lowercase =MAPPING_T2S lowercase =IGNORE_KEYS_T2S elif task == "s2s": lowercase =hf_model.speechta.encoder.prenet.feature_encoder lowercase =MAPPING_S2S lowercase =IGNORE_KEYS_S2S else: raise ValueError(f'Unsupported task: {task}' ) for name, value in fairseq_dict.items(): if should_ignore(lowercase_ , lowercase_ ): logger.info(f'{name} was ignored' ) continue lowercase =False if "conv_layers" in name: load_conv_layer( lowercase_ , lowercase_ , lowercase_ , lowercase_ , hf_model.config.feat_extract_norm == '''group''' , ) lowercase =True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: lowercase , lowercase =key.split('''.*.''' ) if prefix in name and suffix in name: lowercase =suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: lowercase =True if "*" in mapped_key: lowercase =name.split(lowercase_ )[0].split('''.''' )[-2] lowercase =mapped_key.replace('''*''' , lowercase_ ) if "weight_g" in name: lowercase ='''weight_g''' elif "weight_v" in name: lowercase ='''weight_v''' elif "bias" in name: lowercase ='''bias''' elif "weight" in name: lowercase ='''weight''' elif "running_mean" in name: lowercase ='''running_mean''' elif "running_var" in name: lowercase ='''running_var''' elif "num_batches_tracked" in name: lowercase ='''num_batches_tracked''' else: lowercase =None set_recursively(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) continue if not is_used: unused_weights.append(lowercase_ ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[int] ) -> Any: '''simple docstring''' lowercase =full_name.split('''conv_layers.''' )[-1] lowercase =name.split('''.''' ) lowercase =int(items[0] ) lowercase =int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) lowercase =value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) lowercase =value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) lowercase =value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) lowercase =value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(lowercase_ ) @torch.no_grad() def UpperCamelCase ( lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : int=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , ) -> str: '''simple docstring''' if config_path is not None: lowercase =SpeechTaConfig.from_pretrained(lowercase_ ) else: lowercase =SpeechTaConfig() if task == "s2t": lowercase =config.max_text_positions lowercase =SpeechTaForSpeechToText(lowercase_ ) elif task == "t2s": lowercase =1_8_7_6 lowercase =6_0_0 lowercase =config.max_speech_positions lowercase =SpeechTaForTextToSpeech(lowercase_ ) elif task == "s2s": lowercase =1_8_7_6 lowercase =config.max_speech_positions lowercase =SpeechTaForSpeechToSpeech(lowercase_ ) else: raise ValueError(f'Unknown task name: {task}' ) if vocab_path: lowercase =SpeechTaTokenizer(lowercase_ , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it lowercase =AddedToken('''<mask>''' , lstrip=lowercase_ , rstrip=lowercase_ ) lowercase =mask_token tokenizer.add_special_tokens({'''mask_token''': mask_token} ) tokenizer.add_tokens(['''<ctc_blank>'''] ) lowercase =SpeechTaFeatureExtractor() lowercase =SpeechTaProcessor(tokenizer=lowercase_ , feature_extractor=lowercase_ ) processor.save_pretrained(lowercase_ ) lowercase =torch.load(lowercase_ ) recursively_load_weights(fairseq_checkpoint['''model'''] , lowercase_ , lowercase_ ) model.save_pretrained(lowercase_ ) if repo_id: print('''Pushing to the hub...''' ) processor.push_to_hub(lowercase_ ) model.push_to_hub(lowercase_ ) if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( '''--task''', default='''s2t''', type=str, help='''Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--vocab_path''', default=None, type=str, help='''Path to SentencePiece model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) _UpperCAmelCase : Optional[int] = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
72
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
1
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
1
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
72
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
1
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _UpperCAmelCase : Dict = '''bart''' _UpperCAmelCase : Tuple = True @st.cache(allow_output_mutation=lowercase_ ) def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' if LOAD_DENSE_INDEX: lowercase =AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) lowercase =AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) lowercase =qar_model.eval() else: lowercase , lowercase =(None, None) if MODEL_TYPE == "bart": lowercase =AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) lowercase =AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) lowercase =torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) lowercase =sas_model.eval() else: lowercase , lowercase =make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=lowercase_ ) def UpperCamelCase ( ) -> Union[str, Any]: '''simple docstring''' if LOAD_DENSE_INDEX: lowercase =faiss.StandardGpuResources() lowercase =datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] lowercase =np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 1_2_8) , ) lowercase =faiss.IndexFlatIP(1_2_8 ) lowercase =faiss.index_cpu_to_gpu(lowercase_ , 1 , lowercase_ ) wikiaab_gpu_index_flat.add(lowercase_ ) # TODO fix for larger GPU else: lowercase , lowercase =(None, None) lowercase =Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=lowercase_ ) def UpperCamelCase ( ) -> List[str]: '''simple docstring''' lowercase =datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) lowercase =elia['''train_eli5'''] lowercase =np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 1_2_8) ) lowercase =faiss.IndexFlatIP(1_2_8 ) eli5_train_q_index.add(lowercase_ ) return (elia_train, eli5_train_q_index) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Any = load_indexes() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : str = load_models() _UpperCAmelCase , _UpperCAmelCase : Optional[int] = load_train_data() def UpperCamelCase ( lowercase_ : int , lowercase_ : str=1_0 ) -> Dict: '''simple docstring''' lowercase =embed_questions_for_retrieval([question] , lowercase_ , lowercase_ ) lowercase , lowercase =eli5_train_q_index.search(lowercase_ , lowercase_ ) lowercase =[elia_train[int(lowercase_ )] for i in I[0]] return nn_examples def UpperCamelCase ( lowercase_ : Tuple , lowercase_ : Tuple="wiki40b" , lowercase_ : Any="dense" , lowercase_ : int=1_0 ) -> int: '''simple docstring''' if source == "none": lowercase , lowercase =(''' <P> '''.join(['''''' for _ in range(1_1 )] ).strip(), []) else: if method == "dense": lowercase , lowercase =query_qa_dense_index( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: lowercase , lowercase =query_es_index( lowercase_ , lowercase_ , index_name='''english_wiki40b_snippets_100w''' , n_results=lowercase_ , ) lowercase =[ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] lowercase ='''question: {} context: {}'''.format(lowercase_ , lowercase_ ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda lowercase_ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda lowercase_ : None), } ) def UpperCamelCase ( lowercase_ : Dict , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : int=6_4 , lowercase_ : str=2_5_6 , lowercase_ : List[Any]=False , lowercase_ : str=2 , lowercase_ : List[Any]=0.9_5 , lowercase_ : Union[str, Any]=0.8 ) -> Optional[Any]: '''simple docstring''' with torch.no_grad(): lowercase =qa_sas_generate( lowercase_ , lowercase_ , lowercase_ , num_answers=1 , num_beams=lowercase_ , min_len=lowercase_ , max_len=lowercase_ , do_sample=lowercase_ , temp=lowercase_ , top_p=lowercase_ , top_k=lowercase_ , max_input_length=1_0_2_4 , device='''cuda:0''' , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar _UpperCAmelCase : List[str] = '''<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>''' _UpperCAmelCase : List[str] = ''' <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class="img-container"> <!-- Inline parent element --> %s </span> </body> </html> ''' % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _UpperCAmelCase : List[Any] = ''' This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. ''' st.sidebar.markdown(description, unsafe_allow_html=True) _UpperCAmelCase : List[str] = [ '''Answer the question''', '''View the retrieved document only''', '''View the most similar ELI5 question and answer''', '''Show me everything, please!''', ] _UpperCAmelCase : List[str] = st.sidebar.checkbox('''Demo options''') if demo_options: _UpperCAmelCase : Dict = st.sidebar.selectbox( '''''', action_list, index=3, ) _UpperCAmelCase : Optional[Any] = action_list.index(action_st) _UpperCAmelCase : Optional[int] = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) _UpperCAmelCase : Union[str, Any] = show_type == '''Show full text of passages''' else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : List[Any] = True _UpperCAmelCase : str = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: _UpperCAmelCase : Union[str, Any] = ''' ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. ''' st.sidebar.markdown(retriever_info) _UpperCAmelCase : List[str] = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) _UpperCAmelCase : Union[str, Any] = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: _UpperCAmelCase : Dict = '''wiki40b''' _UpperCAmelCase : List[Any] = '''dense''' _UpperCAmelCase : int = '''beam''' _UpperCAmelCase : Optional[Any] = 2 _UpperCAmelCase : Optional[Any] = 64 _UpperCAmelCase : Dict = 2_56 _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Any = None _UpperCAmelCase : Dict = st.sidebar.checkbox('''Generation options''') if generate_options: _UpperCAmelCase : Union[str, Any] = ''' ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder\'s output probabilities. ''' st.sidebar.markdown(generate_info) _UpperCAmelCase : Tuple = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) _UpperCAmelCase : List[Any] = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=2_56, value=64, step=8, format=None, key=None ) _UpperCAmelCase : Tuple = st.sidebar.slider( '''Maximum generation length''', min_value=64, max_value=5_12, value=2_56, step=16, format=None, key=None ) if sampled == "beam": _UpperCAmelCase : Optional[Any] = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _UpperCAmelCase : Any = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _UpperCAmelCase : Union[str, Any] = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _UpperCAmelCase : Any = None # start main text _UpperCAmelCase : int = [ '''<MY QUESTION>''', '''How do people make chocolate?''', '''Why do we get a fever when we are sick?''', '''How can different animals perceive different colors?''', '''What is natural language processing?''', '''What\'s the best way to treat a sunburn?''', '''What exactly are vitamins ?''', '''How does nuclear energy provide electricity?''', '''What\'s the difference between viruses and bacteria?''', '''Why are flutes classified as woodwinds when most of them are made out of metal ?''', '''Why do people like drinking coffee even though it tastes so bad?''', '''What happens when wine ages? How does it make the wine taste better?''', '''If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?''', '''How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?''', '''How does New Zealand have so many large bird predators?''', ] _UpperCAmelCase : Union[str, Any] = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": _UpperCAmelCase : Optional[int] = st.text_input('''Enter your question here:''', '''''') else: _UpperCAmelCase : List[Any] = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": _UpperCAmelCase , _UpperCAmelCase : Any = make_support(question, source=wiki_source, method='''dense''', n_results=10) _UpperCAmelCase , _UpperCAmelCase : List[Any] = make_support(question, source=wiki_source, method='''sparse''', n_results=10) _UpperCAmelCase : Optional[Any] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _UpperCAmelCase : Tuple = support_list[:10] _UpperCAmelCase : int = '''<P> ''' + ''' <P> '''.join([res[-1] for res in support_list]) else: _UpperCAmelCase , _UpperCAmelCase : Dict = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _UpperCAmelCase , _UpperCAmelCase : Tuple = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): _UpperCAmelCase : List[Any] = '''https://en.wikipedia.org/wiki/{}'''.format(res[0].replace(''' ''', '''_''')) _UpperCAmelCase : Optional[int] = res[1].strip() if sec_titles == "": _UpperCAmelCase : Tuple = '''[{}]({})'''.format(res[0], wiki_url) else: _UpperCAmelCase : Dict = sec_titles.split(''' & ''') _UpperCAmelCase : str = ''' & '''.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style="font-family:arial; font-size:10pt;">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: _UpperCAmelCase : Union[str, Any] = find_nearest_training(question) _UpperCAmelCase : str = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) _UpperCAmelCase : List[Any] = [ '''{}. {}'''.format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) _UpperCAmelCase : List[str] = ''' --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* ''' st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
72
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
1
'''simple docstring''' import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _UpperCAmelCase : List[Any] = 16 _UpperCAmelCase : Tuple = 32 def UpperCamelCase ( lowercase_ : Accelerator , lowercase_ : int = 1_6 , lowercase_ : str = "bert-base-cased" ) -> Any: '''simple docstring''' lowercase =AutoTokenizer.from_pretrained(lowercase_ ) lowercase =load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(lowercase_ : Any ): # max_length=None => use the model max length (it's actually the default) lowercase =tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowercase_ , max_length=lowercase_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowercase =datasets.map( lowercase_ , batched=lowercase_ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=lowercase_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase =tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(lowercase_ : Union[str, Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase_ , padding='''max_length''' , max_length=1_2_8 , return_tensors='''pt''' ) return tokenizer.pad(lowercase_ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. lowercase =DataLoader( tokenized_datasets['''train'''] , shuffle=lowercase_ , collate_fn=lowercase_ , batch_size=lowercase_ ) lowercase =DataLoader( tokenized_datasets['''validation'''] , shuffle=lowercase_ , collate_fn=lowercase_ , batch_size=lowercase_ ) return train_dataloader, eval_dataloader def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Optional[int] ) -> Tuple: '''simple docstring''' lowercase =Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase =config['''lr'''] lowercase =int(config['''num_epochs'''] ) lowercase =int(config['''seed'''] ) lowercase =int(config['''batch_size'''] ) lowercase =args.model_name_or_path set_seed(lowercase_ ) lowercase , lowercase =get_dataloaders(lowercase_ , lowercase_ , lowercase_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase =AutoModelForSequenceClassification.from_pretrained(lowercase_ , return_dict=lowercase_ ) # Instantiate optimizer lowercase =( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) lowercase =optimizer_cls(params=model.parameters() , lr=lowercase_ ) if accelerator.state.deepspeed_plugin is not None: lowercase =accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: lowercase =1 lowercase =(len(lowercase_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): lowercase =get_linear_schedule_with_warmup( optimizer=lowercase_ , num_warmup_steps=0 , num_training_steps=lowercase_ , ) else: lowercase =DummyScheduler(lowercase_ , total_num_steps=lowercase_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase , lowercase , lowercase , lowercase , lowercase =accelerator.prepare( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # We need to keep track of how many total steps we have iterated over lowercase =0 # We also need to keep track of the stating epoch so files are named properly lowercase =0 # Now we train the model lowercase =evaluate.load('''glue''' , '''mrpc''' ) lowercase =0 lowercase ={} for epoch in range(lowercase_ , lowercase_ ): model.train() for step, batch in enumerate(lowercase_ ): lowercase =model(**lowercase_ ) lowercase =outputs.loss lowercase =loss / gradient_accumulation_steps accelerator.backward(lowercase_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() lowercase =0 for step, batch in enumerate(lowercase_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase =model(**lowercase_ ) lowercase =outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times lowercase , lowercase =accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase_ ) - 1: lowercase =predictions[: len(eval_dataloader.dataset ) - samples_seen] lowercase =references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase_ , references=lowercase_ , ) lowercase =metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'epoch {epoch}:' , lowercase_ ) lowercase =eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: lowercase =eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( ) -> List[Any]: '''simple docstring''' lowercase =argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=lowercase_ , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=lowercase_ , ) parser.add_argument( '''--output_dir''' , type=lowercase_ , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=lowercase_ , default=lowercase_ , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=lowercase_ , default=3 , help='''Number of train epochs.''' , ) lowercase =parser.parse_args() lowercase ={'''lr''': 2E-5, '''num_epochs''': args.num_epochs, '''seed''': 4_2, '''batch_size''': 1_6} training_function(lowercase_ , lowercase_ ) if __name__ == "__main__": main()
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _UpperCAmelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __magic_name__ : UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Model type selected in the list: ' + ', '.join(__SCREAMING_SNAKE_CASE )} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) UpperCamelCase__ = field( default=1_28 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) UpperCamelCase__ = field( default=1_28 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) UpperCamelCase__ = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) UpperCamelCase__ = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) UpperCamelCase__ = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) UpperCamelCase__ = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) UpperCamelCase__ = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) UpperCamelCase__ = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'train' UpperCamelCase__ = 'dev' class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = 42 UpperCamelCase__ = 42 UpperCamelCase__ = 42 def __init__( self , snake_case_ , snake_case_ , snake_case_ = None , snake_case_ = Split.train , snake_case_ = False , snake_case_ = None , snake_case_ = "pt" , ): lowercase =args lowercase =is_language_sensitive lowercase =SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(snake_case_ , snake_case_ ): try: lowercase =Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) lowercase =mode # Load data features from cache or dataset file lowercase ='''v2''' if args.version_2_with_negative else '''v1''' lowercase =os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase =cached_features_file + '''.lock''' with FileLock(snake_case_ ): if os.path.exists(snake_case_ ) and not args.overwrite_cache: lowercase =time.time() lowercase =torch.load(snake_case_ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowercase =self.old_features['''features'''] lowercase =self.old_features.get('''dataset''' , snake_case_ ) lowercase =self.old_features.get('''examples''' , snake_case_ ) logger.info( f'Loading features from cached file {cached_features_file} [took %.3f s]' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'Deleting cached file {cached_features_file} will allow dataset and examples to be cached in' ''' future run''' ) else: if mode == Split.dev: lowercase =self.processor.get_dev_examples(args.data_dir ) else: lowercase =self.processor.get_train_examples(args.data_dir ) lowercase , lowercase =squad_convert_examples_to_features( examples=self.examples , tokenizer=snake_case_ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=snake_case_ , ) lowercase =time.time() torch.save( {'''features''': self.features, '''dataset''': self.dataset, '''examples''': self.examples} , snake_case_ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]' ) def __len__( self ): return len(self.features ) def __getitem__( self , snake_case_ ): # Convert to Tensors and build dataset lowercase =self.features[i] lowercase =torch.tensor(feature.input_ids , dtype=torch.long ) lowercase =torch.tensor(feature.attention_mask , dtype=torch.long ) lowercase =torch.tensor(feature.token_type_ids , dtype=torch.long ) lowercase =torch.tensor(feature.cls_index , dtype=torch.long ) lowercase =torch.tensor(feature.p_mask , dtype=torch.float ) lowercase =torch.tensor(feature.is_impossible , dtype=torch.float ) lowercase ={ '''input_ids''': input_ids, '''attention_mask''': attention_mask, '''token_type_ids''': token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({'''cls_index''': cls_index, '''p_mask''': p_mask} ) if self.args.version_2_with_negative: inputs.update({'''is_impossible''': is_impossible} ) if self.is_language_sensitive: inputs.update({'''langs''': (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowercase =torch.tensor(feature.start_position , dtype=torch.long ) lowercase =torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({'''start_positions''': start_positions, '''end_positions''': end_positions} ) return inputs
72
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class __magic_name__ : UpperCamelCase__ = LEDConfig UpperCamelCase__ = {} UpperCamelCase__ = 'gelu' def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=False , snake_case_=99 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=20 , snake_case_=2 , snake_case_=1 , snake_case_=0 , snake_case_=4 , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =is_training lowercase =use_labels lowercase =vocab_size lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =max_position_embeddings lowercase =eos_token_id lowercase =pad_token_id lowercase =bos_token_id lowercase =attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after lowercase =self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests lowercase =( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def _A( self ): lowercase =ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowercase =tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowercase =tf.concat([input_ids, eos_tensor] , axis=1 ) lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase =self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) lowercase =prepare_led_inputs_dict(snake_case_ , snake_case_ , snake_case_ ) lowercase =tf.concat( [tf.zeros_like(snake_case_ )[:, :-1], tf.ones_like(snake_case_ )[:, -1:]] , axis=-1 , ) lowercase =global_attention_mask return config, inputs_dict def _A( self , snake_case_ , snake_case_ ): lowercase =TFLEDModel(config=snake_case_ ).get_decoder() lowercase =inputs_dict['''input_ids'''] lowercase =input_ids[:1, :] lowercase =inputs_dict['''attention_mask'''][:1, :] lowercase =1 # first forward pass lowercase =model(snake_case_ , attention_mask=snake_case_ , use_cache=snake_case_ ) lowercase , lowercase =outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase =ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase =tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowercase =tf.concat([input_ids, next_tokens] , axis=-1 ) lowercase =tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowercase =model(snake_case_ , attention_mask=snake_case_ )[0] lowercase =model(snake_case_ , attention_mask=snake_case_ , past_key_values=snake_case_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowercase =int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowercase =output_from_no_past[:, -3:, random_slice_idx] lowercase =output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(snake_case_ , snake_case_ , rtol=1E-3 ) def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : Optional[int]=None , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[int]=None , lowercase_ : str=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowercase =tf.cast(tf.math.not_equal(lowercase_ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowercase =tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowercase =tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowercase =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () UpperCamelCase__ = (TFLEDForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =TFLEDModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =tf.zeros_like(inputs_dict['''attention_mask'''] ) lowercase =2 lowercase =tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['''global_attention_mask'''] , ) lowercase =True lowercase =self.model_tester.seq_length lowercase =self.model_tester.encoder_seq_length def check_decoder_attentions_output(snake_case_ ): lowercase =outputs.decoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(snake_case_ ): lowercase =[t.numpy() for t in outputs.encoder_attentions] lowercase =[t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: lowercase =True lowercase =False lowercase =False lowercase =model_class(snake_case_ ) lowercase =model(self._prepare_for_class(snake_case_ , snake_case_ ) ) lowercase =len(snake_case_ ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) if self.is_encoder_decoder: lowercase =model_class(snake_case_ ) lowercase =model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_decoder_attentions_output(snake_case_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowercase =True lowercase =model_class(snake_case_ ) lowercase =model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) # Check attention is always last and order is fine lowercase =True lowercase =True lowercase =model_class(snake_case_ ) lowercase =model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(snake_case_ ) ) self.assertEqual(model.config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) @unittest.skip('''LED keeps using potentially symbolic tensors in conditionals and breaks tracing.''' ) def _A( self ): pass def _A( self ): # TODO: Head-masking not yet implement pass def UpperCamelCase ( lowercase_ : int ) -> Optional[int]: '''simple docstring''' return tf.constant(lowercase_ , dtype=tf.intaa ) _UpperCAmelCase : List[Any] = 1e-4 @slow @require_tf class __magic_name__ ( unittest.TestCase ): def _A( self ): lowercase =TFLEDForConditionalGeneration.from_pretrained('''allenai/led-base-16384''' ).led # change to intended input here lowercase =_long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) lowercase =_long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) lowercase =prepare_led_inputs_dict(model.config , snake_case_ , snake_case_ ) lowercase =model(**snake_case_ )[0] lowercase =(1, 10_24, 7_68) self.assertEqual(output.shape , snake_case_ ) # change to expected output here lowercase =tf.convert_to_tensor( [[2.30_50, 2.82_79, 0.65_31], [-1.84_57, -0.14_55, -3.56_61], [-1.01_86, 0.45_86, -2.20_43]] , ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-3 ) def _A( self ): lowercase =TFLEDForConditionalGeneration.from_pretrained('''allenai/led-base-16384''' ) # change to intended input here lowercase =_long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) lowercase =_long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) lowercase =prepare_led_inputs_dict(model.config , snake_case_ , snake_case_ ) lowercase =model(**snake_case_ )[0] lowercase =(1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , snake_case_ ) # change to expected output here lowercase =tf.convert_to_tensor( [[33.65_07, 6.45_72, 16.80_89], [5.87_39, -2.42_38, 11.29_02], [-3.21_39, -4.31_49, 4.27_83]] , ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-3 , rtol=1E-3 )
72
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py _UpperCAmelCase : List[Any] = '''.''' if __name__ == "__main__": _UpperCAmelCase : int = os.path.join(REPO_PATH, '''utils/documentation_tests.txt''') _UpperCAmelCase : str = [] _UpperCAmelCase : Optional[Any] = [] with open(doctest_file_path) as fp: for line in fp: _UpperCAmelCase : str = line.strip() _UpperCAmelCase : Tuple = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: _UpperCAmelCase : Tuple = '''\n'''.join(non_existent_paths) raise ValueError(F"""`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}""") if all_paths != sorted(all_paths): raise ValueError('''Files in `utils/documentation_tests.txt` are not in alphabetical order.''')
72
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
1
'''simple docstring''' import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=False , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =is_training lowercase =use_input_mask lowercase =use_token_type_ids lowercase =use_labels lowercase =vocab_size lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =max_position_embeddings lowercase =type_vocab_size lowercase =type_sequence_label_size lowercase =initializer_range lowercase =num_labels lowercase =num_choices lowercase =scope def _A( self ): lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase =None if self.use_input_mask: lowercase =random_attention_mask([self.batch_size, self.seq_length] ) lowercase =None if self.use_token_type_ids: lowercase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase =None lowercase =None lowercase =None if self.use_labels: lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase =ids_tensor([self.batch_size] , self.num_choices ) lowercase =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _A( self ): return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case_ , initializer_range=self.initializer_range , use_stable_embedding=snake_case_ , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =OpenLlamaModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , attention_mask=snake_case_ ) lowercase =model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =True lowercase =OpenLlamaModel(snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , ) lowercase =model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , ) lowercase =model(snake_case_ , attention_mask=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =OpenLlamaForCausalLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =True lowercase =True lowercase =OpenLlamaForCausalLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() # first forward pass lowercase =model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , use_cache=snake_case_ , ) lowercase =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase =ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowercase =torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase =torch.cat([input_mask, next_mask] , dim=-1 ) lowercase =model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , output_hidden_states=snake_case_ , )['''hidden_states'''][0] lowercase =model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , past_key_values=snake_case_ , output_hidden_states=snake_case_ , )['''hidden_states'''][0] # select random slice lowercase =ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase =output_from_no_past[:, -3:, random_slice_idx].detach() lowercase =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case_ , snake_case_ , atol=1E-3 ) ) def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) UpperCamelCase__ = (OpenLlamaForCausalLM,) if is_torch_available() else () UpperCamelCase__ = ( { 'feature-extraction': OpenLlamaModel, 'text-classification': OpenLlamaForSequenceClassification, 'text-generation': OpenLlamaForCausalLM, 'zero-shot': OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =OpenLlamaModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase =type self.model_tester.create_and_check_model(*snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =3 lowercase =input_dict['''input_ids'''] lowercase =input_ids.ne(1 ).to(snake_case_ ) lowercase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase =OpenLlamaForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =3 lowercase ='''single_label_classification''' lowercase =input_dict['''input_ids'''] lowercase =input_ids.ne(1 ).to(snake_case_ ) lowercase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase =OpenLlamaForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =3 lowercase ='''multi_label_classification''' lowercase =input_dict['''input_ids'''] lowercase =input_ids.ne(1 ).to(snake_case_ ) lowercase =ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowercase =OpenLlamaForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('''Open-Llama buffers include complex numbers, which breaks this test''' ) def _A( self ): pass @parameterized.expand([('''linear''',), ('''dynamic''',)] ) def _A( self , snake_case_ ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =ids_tensor([1, 10] , config.vocab_size ) lowercase =ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase =OpenLlamaModel(snake_case_ ) original_model.to(snake_case_ ) original_model.eval() lowercase =original_model(snake_case_ ).last_hidden_state lowercase =original_model(snake_case_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase ={'''type''': scaling_type, '''factor''': 10.0} lowercase =OpenLlamaModel(snake_case_ ) scaled_model.to(snake_case_ ) scaled_model.eval() lowercase =scaled_model(snake_case_ ).last_hidden_state lowercase =scaled_model(snake_case_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case_ , snake_case_ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(snake_case_ , snake_case_ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case_ , snake_case_ , atol=1E-5 ) )
72
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
1
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def UpperCamelCase ( lowercase_ : str , lowercase_ : str , lowercase_ : str , lowercase_ : PreTrainedTokenizer , lowercase_ : int , lowercase_ : Optional[int] = None , ) -> int: '''simple docstring''' lowercase ={} if train_file is not None: lowercase =[train_file] if eval_file is not None: lowercase =[eval_file] if test_file is not None: lowercase =[test_file] lowercase =datasets.load_dataset('''csv''' , data_files=lowercase_ ) lowercase =list(ds[list(files.keys() )[0]].features.keys() ) lowercase =features_name.pop(lowercase_ ) lowercase =list(set(ds[list(files.keys() )[0]][label_name] ) ) lowercase ={label: i for i, label in enumerate(lowercase_ )} lowercase =tokenizer.model_input_names lowercase ={} if len(lowercase_ ) == 1: for k in files.keys(): lowercase =ds[k].map( lambda lowercase_ : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=lowercase_ , max_length=lowercase_ , padding='''max_length''' ) , batched=lowercase_ , ) elif len(lowercase_ ) == 2: for k in files.keys(): lowercase =ds[k].map( lambda lowercase_ : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=lowercase_ , max_length=lowercase_ , padding='''max_length''' , ) , batched=lowercase_ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: lowercase ={k: v for k, v in ex.items() if k in input_names} lowercase =labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: lowercase ={k: v for k, v in ex.items() if k in input_names} lowercase =labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: lowercase ={k: v for k, v in ex.items() if k in input_names} lowercase =labelaid[ex[label_name]] yield (d, label) lowercase =( tf.data.Dataset.from_generator( lowercase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: lowercase =train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) lowercase =( tf.data.Dataset.from_generator( lowercase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: lowercase =val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) lowercase =( tf.data.Dataset.from_generator( lowercase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: lowercase =test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid _UpperCAmelCase : List[Any] = logging.getLogger(__name__) @dataclass class __magic_name__ : UpperCamelCase__ = field(metadata={'help': 'Which column contains the label'} ) UpperCamelCase__ = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The path of the training file'} ) UpperCamelCase__ = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The path of the development file'} ) UpperCamelCase__ = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'The path of the test file'} ) UpperCamelCase__ = field( default=1_28 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) @dataclass class __magic_name__ : UpperCamelCase__ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) UpperCamelCase__ = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Set this flag to use fast tokenization.'} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. UpperCamelCase__ = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) def UpperCamelCase ( ) -> Optional[int]: '''simple docstring''' lowercase =HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) lowercase , lowercase , lowercase =parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. Use' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info( f'n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, ' f'16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase =AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowercase , lowercase , lowercase , lowercase =get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=lowercase_ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) lowercase =AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(lowercase_ ) , labelaid=lowercase_ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='''text-classification''' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): lowercase =TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('''.bin''' in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , ) def compute_metrics(lowercase_ : EvalPrediction ) -> Dict: lowercase =np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer lowercase =TFTrainer( model=lowercase_ , args=lowercase_ , train_dataset=lowercase_ , eval_dataset=lowercase_ , compute_metrics=lowercase_ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase ={} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase =trainer.evaluate() lowercase =os.path.join(training_args.output_dir , '''eval_results.txt''' ) with open(lowercase_ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(f' {key} = {value}' ) writer.write(f'{key} = {value}\n' ) results.update(lowercase_ ) return results if __name__ == "__main__": main()
72
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
1
'''simple docstring''' import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class __magic_name__ ( unittest.TestCase ): def __init__( self , snake_case_ , snake_case_=1_00 , snake_case_=13 , snake_case_=30 , snake_case_=2 , snake_case_=3 , snake_case_=True , snake_case_=True , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=10 , snake_case_=0.02 , snake_case_=3 , ): lowercase =parent lowercase =vocab_size lowercase =batch_size lowercase =image_size lowercase =patch_size lowercase =num_channels lowercase =is_training lowercase =use_labels lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =type_sequence_label_size lowercase =initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowercase =(image_size // patch_size) ** 2 lowercase =num_patches + 1 def _A( self ): lowercase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase =None if self.use_labels: lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase =BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) return config, pixel_values, labels def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =FlaxBeitModel(config=snake_case_ ) lowercase =model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =FlaxBeitForMaskedImageModeling(config=snake_case_ ) lowercase =model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =self.type_sequence_label_size lowercase =FlaxBeitForImageClassification(config=snake_case_ ) lowercase =model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowercase =1 lowercase =FlaxBeitForImageClassification(snake_case_ ) lowercase =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase =model(snake_case_ ) def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def _A( self ): lowercase =FlaxBeitModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =['''pixel_values'''] self.assertListEqual(arg_names[:1] , snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase =self._prepare_for_class(snake_case_ , snake_case_ ) lowercase =model_class(snake_case_ ) @jax.jit def model_jitted(snake_case_ , **snake_case_ ): return model(pixel_values=snake_case_ , **snake_case_ ) with self.subTest('''JIT Enabled''' ): lowercase =model_jitted(**snake_case_ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): lowercase =model_jitted(**snake_case_ ).to_tuple() self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for jitted_output, output in zip(snake_case_ , snake_case_ ): self.assertEqual(jitted_output.shape , output.shape ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def _A( self ): for model_class_name in self.all_model_classes: lowercase =model_class_name.from_pretrained('''microsoft/beit-base-patch16-224''' ) lowercase =model(np.ones((1, 3, 2_24, 2_24) ) ) self.assertIsNotNone(snake_case_ ) def UpperCamelCase ( ) -> str: '''simple docstring''' lowercase =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @require_flax class __magic_name__ ( unittest.TestCase ): @cached_property def _A( self ): return BeitImageProcessor.from_pretrained('''microsoft/beit-base-patch16-224''' ) if is_vision_available() else None @slow def _A( self ): lowercase =FlaxBeitForMaskedImageModeling.from_pretrained('''microsoft/beit-base-patch16-224-pt22k''' ) lowercase =self.default_image_processor lowercase =prepare_img() lowercase =image_processor(images=snake_case_ , return_tensors='''np''' ).pixel_values # prepare bool_masked_pos lowercase =np.ones((1, 1_96) , dtype=snake_case_ ) # forward pass lowercase =model(pixel_values=snake_case_ , bool_masked_pos=snake_case_ ) lowercase =outputs.logits # verify the logits lowercase =(1, 1_96, 81_92) self.assertEqual(logits.shape , snake_case_ ) lowercase =np.array( [[-3.24_37, 0.50_72, -13.91_74], [-3.24_56, 0.49_48, -13.94_01], [-3.20_33, 0.51_21, -13.85_50]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , snake_case_ , atol=1E-2 ) ) @slow def _A( self ): lowercase =FlaxBeitForImageClassification.from_pretrained('''microsoft/beit-base-patch16-224''' ) lowercase =self.default_image_processor lowercase =prepare_img() lowercase =image_processor(images=snake_case_ , return_tensors='''np''' ) # forward pass lowercase =model(**snake_case_ ) lowercase =outputs.logits # verify the logits lowercase =(1, 10_00) self.assertEqual(logits.shape , snake_case_ ) lowercase =np.array([-1.23_85, -1.09_87, -1.01_08] ) self.assertTrue(np.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) ) lowercase =2_81 self.assertEqual(logits.argmax(-1 ).item() , snake_case_ ) @slow def _A( self ): lowercase =FlaxBeitForImageClassification.from_pretrained('''microsoft/beit-large-patch16-224-pt22k-ft22k''' ) lowercase =self.default_image_processor lowercase =prepare_img() lowercase =image_processor(images=snake_case_ , return_tensors='''np''' ) # forward pass lowercase =model(**snake_case_ ) lowercase =outputs.logits # verify the logits lowercase =(1, 2_18_41) self.assertEqual(logits.shape , snake_case_ ) lowercase =np.array([1.68_81, -0.27_87, 0.59_01] ) self.assertTrue(np.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) ) lowercase =23_96 self.assertEqual(logits.argmax(-1 ).item() , snake_case_ )
72
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
1
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def UpperCamelCase ( lowercase_ : int ) -> Tuple: '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @staticmethod def _A( snake_case_ ): lowercase =parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=snake_case_ , default=snake_case_ , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=snake_case_ , help='''Name of the model to download''' ) download_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =model lowercase =cache lowercase =force lowercase =trust_remote_code def _A( self ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
72
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
1
'''simple docstring''' import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=[30, 30] , snake_case_=2 , snake_case_=3 , snake_case_=True , snake_case_=True , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=10 , snake_case_=0.02 , snake_case_=3 , snake_case_=None , snake_case_=8 , snake_case_=10 , ): lowercase =parent lowercase =batch_size lowercase =image_size lowercase =patch_size lowercase =num_channels lowercase =is_training lowercase =use_labels lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =type_sequence_label_size lowercase =initializer_range lowercase =num_labels lowercase =scope lowercase =n_targets lowercase =num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens lowercase =(image_size[1] // patch_size) * (image_size[0] // patch_size) lowercase =num_patches + 1 + self.num_detection_tokens def _A( self ): lowercase =floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) lowercase =None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) lowercase =[] for i in range(self.batch_size ): lowercase ={} lowercase =torch.randint( high=self.num_labels , size=(self.n_targets,) , device=snake_case_ ) lowercase =torch.rand(self.n_targets , 4 , device=snake_case_ ) labels.append(snake_case_ ) lowercase =self.get_config() return config, pixel_values, labels def _A( self ): return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =YolosModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ ): lowercase =YolosForObjectDetection(snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(pixel_values=snake_case_ ) lowercase =model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) lowercase =model(pixel_values=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def _A( self ): lowercase =self.prepare_config_and_inputs() lowercase , lowercase , lowercase =config_and_inputs lowercase ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (YolosModel, YolosForObjectDetection) if is_torch_available() else () UpperCamelCase__ = ( {'feature-extraction': YolosModel, 'object-detection': YolosForObjectDetection} if is_torch_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self , snake_case_ , snake_case_ , snake_case_=False ): lowercase =super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": lowercase =[] for i in range(self.model_tester.batch_size ): lowercase ={} lowercase =torch.ones( size=(self.model_tester.n_targets,) , device=snake_case_ , dtype=torch.long ) lowercase =torch.ones( self.model_tester.n_targets , 4 , device=snake_case_ , dtype=torch.float ) labels.append(snake_case_ ) lowercase =labels return inputs_dict def _A( self ): lowercase =YolosModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): # YOLOS does not use inputs_embeds pass def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase =model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =['''pixel_values'''] self.assertListEqual(arg_names[:1] , snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() lowercase =True # in YOLOS, the seq_len is different lowercase =self.model_tester.expected_seq_len for model_class in self.all_model_classes: lowercase =True lowercase =False lowercase =True lowercase =model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): lowercase =model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) lowercase =outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowercase =True lowercase =model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): lowercase =model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) lowercase =outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) lowercase =len(snake_case_ ) # Check attention is always last and order is fine lowercase =True lowercase =True lowercase =model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): lowercase =model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) lowercase =1 self.assertEqual(out_len + added_hidden_states , len(snake_case_ ) ) lowercase =outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def _A( self ): def check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ): lowercase =model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): lowercase =model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) lowercase =outputs.hidden_states lowercase =getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) # YOLOS has a different seq_length lowercase =self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase =True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*snake_case_ ) @slow def _A( self ): for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =YolosModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def UpperCamelCase ( ) -> Tuple: '''simple docstring''' lowercase =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __magic_name__ ( unittest.TestCase ): @cached_property def _A( self ): return AutoImageProcessor.from_pretrained('''hustvl/yolos-small''' ) if is_vision_available() else None @slow def _A( self ): lowercase =YolosForObjectDetection.from_pretrained('''hustvl/yolos-small''' ).to(snake_case_ ) lowercase =self.default_image_processor lowercase =prepare_img() lowercase =image_processor(images=snake_case_ , return_tensors='''pt''' ).to(snake_case_ ) # forward pass with torch.no_grad(): lowercase =model(inputs.pixel_values ) # verify outputs lowercase =torch.Size((1, 1_00, 92) ) self.assertEqual(outputs.logits.shape , snake_case_ ) lowercase =torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] , device=snake_case_ , ) lowercase =torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] , device=snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , snake_case_ , atol=1E-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , snake_case_ , atol=1E-4 ) ) # verify postprocessing lowercase =image_processor.post_process_object_detection( snake_case_ , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] lowercase =torch.tensor([0.99_94, 0.97_90, 0.99_64, 0.99_72, 0.98_61] ).to(snake_case_ ) lowercase =[75, 75, 17, 63, 17] lowercase =torch.tensor([3_35.06_09, 79.38_48, 3_75.42_16, 1_87.24_95] ).to(snake_case_ ) self.assertEqual(len(results['''scores'''] ) , 5 ) self.assertTrue(torch.allclose(results['''scores'''] , snake_case_ , atol=1E-4 ) ) self.assertSequenceEqual(results['''labels'''].tolist() , snake_case_ ) self.assertTrue(torch.allclose(results['''boxes'''][0, :] , snake_case_ ) )
72
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : List[Any] = {'''vocab_file''': '''sentencepiece.bpe.model'''} _UpperCAmelCase : Optional[int] = { '''vocab_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model''' ), }, } _UpperCAmelCase : Tuple = { '''moussaKam/mbarthez''': 10_24, '''moussaKam/barthez''': 10_24, '''moussaKam/barthez-orangesum-title''': 10_24, } _UpperCAmelCase : Any = '''▁''' class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<s>" , snake_case_="<unk>" , snake_case_="<pad>" , snake_case_="<mask>" , snake_case_ = None , **snake_case_ , ): # Mask token behave like a normal word, i.e. include the space before it lowercase =AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , cls_token=snake_case_ , pad_token=snake_case_ , mask_token=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(snake_case_ ) ) lowercase ={'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} lowercase =len(self.sp_model ) - 1 lowercase ={v: k for k, v in self.fairseq_tokens_to_ids.items()} def _A( self , snake_case_ , snake_case_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase =[self.cls_token_id] lowercase =[self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) if token_ids_a is None: return [1] + ([0] * len(snake_case_ )) + [1] return [1] + ([0] * len(snake_case_ )) + [1, 1] + ([0] * len(snake_case_ )) + [1] def _A( self , snake_case_ , snake_case_ = None ): lowercase =[self.sep_token_id] lowercase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _A( self ): return len(self.sp_model ) def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase =self.sp_model.PieceToId(snake_case_ ) return spm_id if spm_id else self.unk_token_id def _A( self , snake_case_ ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(snake_case_ ) def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' lowercase =False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case_ ) + token lowercase =True lowercase =[] else: current_sub_tokens.append(snake_case_ ) lowercase =False out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
1
'''simple docstring''' import math import os import sys def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' lowercase ='''''' try: with open(lowercase_ , '''rb''' ) as binary_file: lowercase =binary_file.read() for dat in data: lowercase =f'{dat:08b}' result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def UpperCamelCase ( lowercase_ : dict[str, str] , lowercase_ : str , lowercase_ : int , lowercase_ : str ) -> None: '''simple docstring''' lexicon.pop(lowercase_ ) lowercase =last_match_id if math.loga(lowercase_ ).is_integer(): for curr_key in lexicon: lowercase ='''0''' + lexicon[curr_key] lowercase =bin(lowercase_ )[2:] def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' lowercase ={'''0''': '''0''', '''1''': '''1'''} lowercase , lowercase ='''''', '''''' lowercase =len(lowercase_ ) for i in range(len(lowercase_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue lowercase =lexicon[curr_string] result += last_match_id add_key_to_lexicon(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) index += 1 lowercase ='''''' while curr_string != "" and curr_string not in lexicon: curr_string += "0" if curr_string != "": lowercase =lexicon[curr_string] result += last_match_id return result def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> str: '''simple docstring''' lowercase =os.path.getsize(lowercase_ ) lowercase =bin(lowercase_ )[2:] lowercase =len(lowercase_ ) return "0" * (length_length - 1) + file_length_binary + compressed def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> None: '''simple docstring''' lowercase =8 try: with open(lowercase_ , '''wb''' ) as opened_file: lowercase =[ to_write[i : i + byte_length] for i in range(0 , len(lowercase_ ) , lowercase_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array: opened_file.write(int(lowercase_ , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> None: '''simple docstring''' lowercase =read_file_binary(lowercase_ ) lowercase =compress_data(lowercase_ ) lowercase =add_file_length(lowercase_ , lowercase_ ) write_file_binary(lowercase_ , lowercase_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
72
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
1
'''simple docstring''' import unittest import torch from torch import nn from diffusers.models.activations import get_activation class __magic_name__ ( unittest.TestCase ): def _A( self ): lowercase =get_activation('''swish''' ) self.assertIsInstance(snake_case_ , nn.SiLU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def _A( self ): lowercase =get_activation('''silu''' ) self.assertIsInstance(snake_case_ , nn.SiLU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def _A( self ): lowercase =get_activation('''mish''' ) self.assertIsInstance(snake_case_ , nn.Mish ) self.assertEqual(act(torch.tensor(-2_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def _A( self ): lowercase =get_activation('''gelu''' ) self.assertIsInstance(snake_case_ , nn.GELU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 )
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BlenderbotSmallTokenizer UpperCamelCase__ = False def _A( self ): super().setUp() lowercase =['''__start__''', '''adapt''', '''act''', '''ap@@''', '''te''', '''__end__''', '''__unk__'''] lowercase =dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) lowercase =['''#version: 0.2''', '''a p''', '''t e</w>''', '''ap t</w>''', '''a d''', '''ad apt</w>''', '''a c''', '''ac t</w>''', ''''''] lowercase ={'''unk_token''': '''__unk__''', '''bos_token''': '''__start__''', '''eos_token''': '''__end__'''} lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(snake_case_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(snake_case_ ) ) def _A( self , **snake_case_ ): kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def _A( self , snake_case_ ): lowercase ='''adapt act apte''' lowercase ='''adapt act apte''' return input_text, output_text def _A( self ): lowercase =BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowercase ='''adapt act apte''' lowercase =['''adapt''', '''act''', '''ap@@''', '''te'''] lowercase =tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =[tokenizer.bos_token] + tokens + [tokenizer.eos_token] lowercase =[0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) assert tok('''sam''' ).input_ids == [13_84] lowercase ='''I am a small frog.''' lowercase =tok([src_text] , padding=snake_case_ , truncation=snake_case_ )['''input_ids'''] lowercase =tok.batch_decode(snake_case_ , skip_special_tokens=snake_case_ , clean_up_tokenization_spaces=snake_case_ )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def _A( self ): lowercase =BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) lowercase ='''I am a small frog .''' lowercase ='''.''' lowercase =tok(snake_case_ )['''input_ids'''] lowercase =tok(snake_case_ )['''input_ids'''] assert encoded[-1] == encoded_dot[0]
72
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
1
'''simple docstring''' import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def UpperCamelCase ( lowercase_ : List[str] ) -> int: '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =np.max(_outputs , axis=-1 , keepdims=lowercase_ ) lowercase =np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=lowercase_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'sigmoid' UpperCamelCase__ = 'softmax' UpperCamelCase__ = 'none' @add_end_docstrings( __SCREAMING_SNAKE_CASE , r'\n return_all_scores (`bool`, *optional*, defaults to `False`):\n Whether to return all prediction scores or just the one of the predicted class.\n function_to_apply (`str`, *optional*, defaults to `"default"`):\n The function to apply to the model outputs in order to retrieve the scores. Accepts four different values:\n\n - `"default"`: if the model has a single label, will apply the sigmoid function on the output. If the model\n has several labels, will apply the softmax function on the output.\n - `"sigmoid"`: Applies the sigmoid function on the output.\n - `"softmax"`: Applies the softmax function on the output.\n - `"none"`: Does not apply any function on the output.\n ' , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = False UpperCamelCase__ = ClassificationFunction.NONE def __init__( self , **snake_case_ ): super().__init__(**snake_case_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def _A( self , snake_case_=None , snake_case_=None , snake_case_="" , **snake_case_ ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" lowercase =tokenizer_kwargs lowercase ={} if hasattr(self.model.config , '''return_all_scores''' ) and return_all_scores is None: lowercase =self.model.config.return_all_scores if isinstance(snake_case_ , snake_case_ ) or top_k is None: lowercase =top_k lowercase =False elif return_all_scores is not None: warnings.warn( '''`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of''' ''' `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.''' , snake_case_ , ) if return_all_scores: lowercase =None else: lowercase =1 if isinstance(snake_case_ , snake_case_ ): lowercase =ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: lowercase =function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *snake_case_ , **snake_case_ ): lowercase =super().__call__(*snake_case_ , **snake_case_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. lowercase ='''top_k''' not in kwargs if isinstance(args[0] , snake_case_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def _A( self , snake_case_ , **snake_case_ ): lowercase =self.framework if isinstance(snake_case_ , snake_case_ ): return self.tokenizer(**snake_case_ , return_tensors=snake_case_ , **snake_case_ ) elif isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1 and isinstance(inputs[0] , snake_case_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=snake_case_ , **snake_case_ ) elif isinstance(snake_case_ , snake_case_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( '''The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a''' ''' dictionary `{"text": "My text", "text_pair": "My pair"}` in order to send a text pair.''' ) return self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) def _A( self , snake_case_ ): return self.model(**snake_case_ ) def _A( self , snake_case_ , snake_case_=None , snake_case_=1 , snake_case_=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: lowercase =ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: lowercase =ClassificationFunction.SOFTMAX elif hasattr(self.model.config , '''function_to_apply''' ) and function_to_apply is None: lowercase =self.model.config.function_to_apply else: lowercase =ClassificationFunction.NONE lowercase =model_outputs['''logits'''][0] lowercase =outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: lowercase =sigmoid(snake_case_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: lowercase =softmax(snake_case_ ) elif function_to_apply == ClassificationFunction.NONE: lowercase =outputs else: raise ValueError(f'Unrecognized `function_to_apply` argument: {function_to_apply}' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} lowercase =[ {'''label''': self.model.config.idalabel[i], '''score''': score.item()} for i, score in enumerate(snake_case_ ) ] if not _legacy: dict_scores.sort(key=lambda snake_case_ : x["score"] , reverse=snake_case_ ) if top_k is not None: lowercase =dict_scores[:top_k] return dict_scores
72
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
1