code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __snake_case ( __lowerCamelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , A : str , A : Tuple=13 , A : Dict=7 , A : Union[str, Any]=True , A : str=True , A : Union[str, Any]=False , A : Any=True , A : List[Any]=99 , A : Optional[Any]=32 , A : List[str]=5 , A : List[str]=4 , A : Tuple=37 , A : Dict="gelu" , A : Optional[int]=0.1 , A : List[str]=0.1 , A : str=512 , A : Dict=16 , A : List[Any]=2 , A : int=0.02 , A : Optional[int]=3 , A : int=4 , A : List[str]=None , ): __snake_case: Any = parent __snake_case: Tuple = batch_size __snake_case: List[Any] = seq_length __snake_case: Dict = is_training __snake_case: Optional[int] = use_input_mask __snake_case: int = use_token_type_ids __snake_case: Dict = use_labels __snake_case: int = vocab_size __snake_case: List[Any] = hidden_size __snake_case: List[str] = num_hidden_layers __snake_case: List[str] = num_attention_heads __snake_case: Dict = intermediate_size __snake_case: str = hidden_act __snake_case: List[Any] = hidden_dropout_prob __snake_case: List[str] = attention_probs_dropout_prob __snake_case: int = max_position_embeddings __snake_case: List[str] = type_vocab_size __snake_case: Any = type_sequence_label_size __snake_case: List[Any] = initializer_range __snake_case: Optional[Any] = num_labels __snake_case: Any = num_choices __snake_case: Tuple = scope def UpperCAmelCase__ ( self : Dict ): __snake_case: Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case: Tuple = None if self.use_input_mask: __snake_case: str = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case: Union[str, Any] = None __snake_case: List[Any] = None __snake_case: str = None if self.use_labels: __snake_case: Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case: Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case: Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) __snake_case: Union[str, Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self : List[Any] ): return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self : Dict , A : List[str] , A : Tuple , A : List[str] , A : Any , A : Dict , A : Optional[Any] ): __snake_case: Any = DistilBertModel(config=A ) model.to(A ) model.eval() __snake_case: str = model(A , A ) __snake_case: List[Any] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase__ ( self : Any , A : Union[str, Any] , A : List[Any] , A : List[Any] , A : Tuple , A : Tuple , A : List[str] ): __snake_case: Union[str, Any] = DistilBertForMaskedLM(config=A ) model.to(A ) model.eval() __snake_case: Optional[Any] = model(A , attention_mask=A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase__ ( self : Union[str, Any] , A : Optional[int] , A : Optional[int] , A : Tuple , A : Union[str, Any] , A : Optional[Any] , A : List[str] ): __snake_case: List[Any] = DistilBertForQuestionAnswering(config=A ) model.to(A ) model.eval() __snake_case: Dict = model( A , attention_mask=A , start_positions=A , end_positions=A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase__ ( self : str , A : Tuple , A : List[str] , A : Optional[Any] , A : Optional[int] , A : List[Any] , A : Any ): __snake_case: Union[str, Any] = self.num_labels __snake_case: str = DistilBertForSequenceClassification(A ) model.to(A ) model.eval() __snake_case: Tuple = model(A , attention_mask=A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Any , A : Tuple , A : Union[str, Any] , A : Optional[int] , A : List[Any] , A : List[Any] , A : Dict ): __snake_case: Dict = self.num_labels __snake_case: List[str] = DistilBertForTokenClassification(config=A ) model.to(A ) model.eval() __snake_case: str = model(A , attention_mask=A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self : Optional[Any] , A : List[Any] , A : str , A : Optional[Any] , A : str , A : List[str] , A : List[Any] ): __snake_case: Any = self.num_choices __snake_case: Union[str, Any] = DistilBertForMultipleChoice(config=A ) model.to(A ) model.eval() __snake_case: Any = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case: Union[str, Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case: Dict = model( A , attention_mask=A , labels=A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: str = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)): Optional[Any] = config_and_inputs __snake_case: Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __snake_case ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) lowerCAmelCase__ = ( { """feature-extraction""": DistilBertModel, """fill-mask""": DistilBertForMaskedLM, """question-answering""": DistilBertForQuestionAnswering, """text-classification""": DistilBertForSequenceClassification, """token-classification""": DistilBertForTokenClassification, """zero-shot""": DistilBertForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = True def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: Union[str, Any] = DistilBertModelTester(self ) __snake_case: Optional[int] = ConfigTester(self , config_class=A , dim=37 ) def UpperCAmelCase__ ( self : Any ): self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : Dict ): __snake_case: Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*A ) def UpperCAmelCase__ ( self : Optional[Any] ): __snake_case: str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*A ) def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*A ) def UpperCAmelCase__ ( self : Dict ): __snake_case: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*A ) def UpperCAmelCase__ ( self : List[Any] ): __snake_case: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*A ) def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*A ) @slow def UpperCAmelCase__ ( self : List[Any] ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case: Tuple = DistilBertModel.from_pretrained(A ) self.assertIsNotNone(A ) @slow @require_torch_gpu def UpperCAmelCase__ ( self : Any ): __snake_case , __snake_case: Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __snake_case: Tuple = True __snake_case: Optional[int] = model_class(config=A ) __snake_case: List[Any] = self._prepare_for_class(A , A ) __snake_case: Dict = torch.jit.trace( A , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(A , os.path.join(A , """traced_model.pt""" ) ) __snake_case: int = torch.jit.load(os.path.join(A , """traced_model.pt""" ) , map_location=A ) loaded(inputs_dict["""input_ids"""].to(A ) , inputs_dict["""attention_mask"""].to(A ) ) @require_torch class __snake_case ( unittest.TestCase ): '''simple docstring''' @slow def UpperCAmelCase__ ( self : str ): __snake_case: Optional[Any] = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __snake_case: Union[str, Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case: Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __snake_case: Any = model(A , attention_mask=A )[0] __snake_case: Dict = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , A ) __snake_case: Optional[int] = torch.tensor( [[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , A , atol=1E-4 ) )
111
from __future__ import annotations import typing from collections import Counter def A__ ( SCREAMING_SNAKE_CASE__) -> typing.Counter[int]: __snake_case: typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1): for perpendicular in range(SCREAMING_SNAKE_CASE__ , max_perimeter + 1): __snake_case: Dict = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(SCREAMING_SNAKE_CASE__): __snake_case: Any = int(base + perpendicular + hypotenuse) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def A__ ( SCREAMING_SNAKE_CASE__ = 1000) -> int: __snake_case: List[str] = pythagorean_triple(SCREAMING_SNAKE_CASE__) return triplets.most_common(1)[0][0] if __name__ == "__main__": print(f'Perimeter {solution()} has maximum solutions')
111
1
from __future__ import annotations def __lowercase ( a__ , a__ , a__ , ) -> tuple: if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('You cannot supply more or less than 2 values' ) elif electron_conc < 0: raise ValueError('Electron concentration cannot be negative in a semiconductor' ) elif hole_conc < 0: raise ValueError('Hole concentration cannot be negative in a semiconductor' ) elif intrinsic_conc < 0: raise ValueError( 'Intrinsic concentration cannot be negative in a semiconductor' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
355
def __lowercase ( a__ ) -> int: __SCREAMING_SNAKE_CASE = [[0 for _ in range(a__ )] for _ in range(m + 1 )] for i in range(m + 1 ): __SCREAMING_SNAKE_CASE = 1 for n in range(m + 1 ): for k in range(1 , a__ ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: lowerCAmelCase__ : Optional[Any] =int(input('''Enter a number: ''').strip()) print(partition(n)) except ValueError: print('''Please enter a number.''') else: try: lowerCAmelCase__ : str =int(sys.argv[1]) print(partition(n)) except ValueError: print('''Please pass a number.''')
118
0
def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = """""" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key _lowerCAmelCase = remove_duplicates(key.upper() ) _lowerCAmelCase = len(snake_case ) # First fill cipher with key characters _lowerCAmelCase = {alphabet[i]: char for i, char in enumerate(snake_case )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(snake_case ) , 26 ): _lowerCAmelCase = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 _lowerCAmelCase = alphabet[i - offset] _lowerCAmelCase = char return cipher_alphabet def _UpperCAmelCase ( snake_case , snake_case ): """simple docstring""" return "".join(cipher_map.get(snake_case , snake_case ) for ch in message.upper() ) def _UpperCAmelCase ( snake_case , snake_case ): """simple docstring""" _lowerCAmelCase = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(snake_case , snake_case ) for ch in message.upper() ) def _UpperCAmelCase ( ): """simple docstring""" _lowerCAmelCase = input("""Enter message to encode or decode: """ ).strip() _lowerCAmelCase = input("""Enter keyword: """ ).strip() _lowerCAmelCase = input("""Encipher or decipher? E/D:""" ).strip()[0].lower() try: _lowerCAmelCase = {"""e""": encipher, """d""": decipher}[option] except KeyError: raise KeyError("""invalid input option""" ) _lowerCAmelCase = create_cipher_map(snake_case ) print(func(snake_case , snake_case ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
82
'''simple docstring''' from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch SCREAMING_SNAKE_CASE_: Optional[int] =logging.get_logger(__name__) @add_end_docstrings( UpperCamelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class __A ( UpperCamelCase__ ): def _lowercase (self : str , __a : GenericTensor ): if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=__a ) else: raise ValueError("Unsupported framework" ) return masked_index def _lowercase (self : Tuple , __a : GenericTensor ): UpperCAmelCase_ = self.get_masked_index(__a ) UpperCAmelCase_ = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( "fill-mask" , self.model.base_model_prefix , f"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def _lowercase (self : List[Any] , __a : GenericTensor ): if isinstance(__a , __a ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input["input_ids"][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(__a ) def _lowercase (self : Tuple , __a : Dict , __a : List[str]=None , **__a : Any ): if return_tensors is None: UpperCAmelCase_ = self.framework UpperCAmelCase_ = self.tokenizer(__a , return_tensors=__a ) self.ensure_exactly_one_mask_token(__a ) return model_inputs def _lowercase (self : str , __a : Optional[int] ): UpperCAmelCase_ = self.model(**__a ) UpperCAmelCase_ = model_inputs["input_ids"] return model_outputs def _lowercase (self : List[str] , __a : Tuple , __a : int=5 , __a : Dict=None ): # Cap top_k if there are targets if target_ids is not None and target_ids.shape[0] < top_k: UpperCAmelCase_ = target_ids.shape[0] UpperCAmelCase_ = model_outputs["input_ids"][0] UpperCAmelCase_ = model_outputs["logits"] if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] UpperCAmelCase_ = outputs.numpy() UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = stable_softmax(__a , axis=-1 ) if target_ids is not None: UpperCAmelCase_ = tf.gather_nd(tf.squeeze(__a , 0 ) , target_ids.reshape(-1 , 1 ) ) UpperCAmelCase_ = tf.expand_dims(__a , 0 ) UpperCAmelCase_ = tf.math.top_k(__a , k=__a ) UpperCAmelCase_ , UpperCAmelCase_ = topk.values.numpy(), topk.indices.numpy() else: UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=__a ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = logits.softmax(dim=-1 ) if target_ids is not None: UpperCAmelCase_ = probs[..., target_ids] UpperCAmelCase_ , UpperCAmelCase_ = probs.topk(__a ) UpperCAmelCase_ = [] UpperCAmelCase_ = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): UpperCAmelCase_ = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place UpperCAmelCase_ = input_ids.numpy().copy() if target_ids is not None: UpperCAmelCase_ = target_ids[p].tolist() UpperCAmelCase_ = p # Filter padding out: UpperCAmelCase_ = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back UpperCAmelCase_ = self.tokenizer.decode(__a , skip_special_tokens=__a ) UpperCAmelCase_ = {"score": v, "token": p, "token_str": self.tokenizer.decode([p] ), "sequence": sequence} row.append(__a ) result.append(__a ) if single_mask: return result[0] return result def _lowercase (self : Dict , __a : List[Any] , __a : List[str]=None ): if isinstance(__a , __a ): UpperCAmelCase_ = [targets] try: UpperCAmelCase_ = self.tokenizer.get_vocab() except Exception: UpperCAmelCase_ = {} UpperCAmelCase_ = [] for target in targets: UpperCAmelCase_ = vocab.get(__a , __a ) if id_ is None: UpperCAmelCase_ = self.tokenizer( __a , add_special_tokens=__a , return_attention_mask=__a , return_token_type_ids=__a , max_length=1 , truncation=__a , )["input_ids"] if len(__a ) == 0: logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ "We cannot replace it with anything meaningful, ignoring it" ) continue UpperCAmelCase_ = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ f"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) UpperCAmelCase_ = list(set(__a ) ) if len(__a ) == 0: raise ValueError("At least one target must be provided when passed." ) UpperCAmelCase_ = np.array(__a ) return target_ids def _lowercase (self : Tuple , __a : Dict=None , __a : List[str]=None ): UpperCAmelCase_ = {} if targets is not None: UpperCAmelCase_ = self.get_target_ids(__a , __a ) UpperCAmelCase_ = target_ids if top_k is not None: UpperCAmelCase_ = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( "fill-mask" , self.model.base_model_prefix , "The tokenizer does not define a `mask_token`." ) return {}, {}, postprocess_params def __call__(self : Union[str, Any] , __a : str , *__a : Any , **__a : Tuple ): UpperCAmelCase_ = super().__call__(__a , **__a ) if isinstance(__a , __a ) and len(__a ) == 1: return outputs[0] return outputs
1
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class lowercase ( UpperCamelCase__ ): _a = "trocr" _a = ["past_key_values"] _a = { "num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model", "num_hidden_layers": "decoder_layers", } def __init__( self , _a=5_0265 , _a=1024 , _a=12 , _a=16 , _a=4096 , _a="gelu" , _a=512 , _a=0.1 , _a=0.0 , _a=0.0 , _a=2 , _a=0.02 , _a=0.0 , _a=True , _a=False , _a=True , _a=True , _a=1 , _a=0 , _a=2 , **_a , ) -> str: _A : List[Any] = vocab_size _A : int = d_model _A : int = decoder_layers _A : Tuple = decoder_attention_heads _A : List[str] = decoder_ffn_dim _A : Tuple = activation_function _A : Dict = max_position_embeddings _A : Any = dropout _A : Union[str, Any] = attention_dropout _A : List[Any] = activation_dropout _A : Tuple = init_std _A : str = decoder_layerdrop _A : Any = use_cache _A : Union[str, Any] = scale_embedding _A : Optional[int] = use_learned_position_embeddings _A : str = layernorm_embedding super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , decoder_start_token_id=_a , **_a , )
343
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class lowercase ( unittest.TestCase ): def a__ ( self ) -> List[str]: debug_launcher(test_script.main ) def a__ ( self ) -> Any: debug_launcher(test_ops.main )
343
1
"""simple docstring""" import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __A = logging.getLogger(__name__) class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' super().__init__( __UpperCAmelCase , question_encoder_tokenizer=__UpperCAmelCase , generator_tokenizer=__UpperCAmelCase , index=__UpperCAmelCase , init_retrieval=__UpperCAmelCase , ) lowerCAmelCase__ :str = None def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' logger.info('initializing retrieval' ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info('dist initialized' ) # needs to be set manually lowerCAmelCase__ :Optional[Any] = self._infer_socket_ifname() # avoid clash with the NCCL port lowerCAmelCase__ :Dict = str(distributed_port + 1 ) lowerCAmelCase__ :List[Any] = dist.new_group(ranks=__UpperCAmelCase , backend='gloo' ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info('dist not initialized / main' ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def snake_case ( self ): '''simple docstring''' return dist.get_rank(group=self.process_group ) == 0 def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=torch.floataa ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = torch.empty(__UpperCAmelCase , dtype=__UpperCAmelCase ) dist.scatter(__UpperCAmelCase , src=0 , scatter_list=__UpperCAmelCase , group=self.process_group ) return target_tensor def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = psutil.net_if_addrs() # a hacky way to deal with varying network interface names lowerCAmelCase__ :Tuple = next((addr for addr in addrs if addr.startswith('e' )) , __UpperCAmelCase ) return ifname def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if not dist.is_initialized(): lowerCAmelCase__ , lowerCAmelCase__ :Union[str, Any] = self._main_retrieve(__UpperCAmelCase , __UpperCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(__UpperCAmelCase ) # distributed training lowerCAmelCase__ :List[str] = dist.get_world_size(group=self.process_group ) # gather logic lowerCAmelCase__ :Any = None if self._is_main(): lowerCAmelCase__ :Dict = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(__UpperCAmelCase )] dist.gather(torch.tensor(__UpperCAmelCase ) , dst=0 , gather_list=__UpperCAmelCase , group=self.process_group ) # scatter logic lowerCAmelCase__ :Optional[int] = question_hidden_states.shape[0] lowerCAmelCase__ :Tuple = [] lowerCAmelCase__ :str = [] if self._is_main(): assert len(__UpperCAmelCase ) == world_size lowerCAmelCase__ , lowerCAmelCase__ :Any = self._main_retrieve(torch.cat(__UpperCAmelCase ).numpy() , __UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ :str = torch.tensor(__UpperCAmelCase ), torch.tensor(__UpperCAmelCase ) lowerCAmelCase__ :Dict = self._chunk_tensor(__UpperCAmelCase , __UpperCAmelCase ) lowerCAmelCase__ :Tuple = self._chunk_tensor(__UpperCAmelCase , __UpperCAmelCase ) lowerCAmelCase__ :List[str] = self._scattered(__UpperCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) lowerCAmelCase__ :Optional[int] = self._scattered(__UpperCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(__UpperCAmelCase )
293
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask __A = logging.getLogger(__name__) class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self , __UpperCAmelCase=-1 ): '''simple docstring''' lowerCAmelCase__ :Dict = label_idx def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if isinstance(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :Optional[Any] = mode.value lowerCAmelCase__ :List[str] = os.path.join(__UpperCAmelCase , F"{mode}.txt" ) lowerCAmelCase__ :List[str] = 1 lowerCAmelCase__ :Union[str, Any] = [] with open(__UpperCAmelCase , encoding='utf-8' ) as f: lowerCAmelCase__ :str = [] lowerCAmelCase__ :Dict = [] for line in f: if line.startswith('-DOCSTART-' ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=F"{mode}-{guid_index}" , words=__UpperCAmelCase , labels=__UpperCAmelCase ) ) guid_index += 1 lowerCAmelCase__ :Tuple = [] lowerCAmelCase__ :List[str] = [] else: lowerCAmelCase__ :List[str] = line.split(' ' ) words.append(splits[0] ) if len(__UpperCAmelCase ) > 1: labels.append(splits[self.label_idx].replace('\n' , '' ) ) else: # Examples could have no label for mode = "test" labels.append('O' ) if words: examples.append(InputExample(guid=F"{mode}-{guid_index}" , words=__UpperCAmelCase , labels=__UpperCAmelCase ) ) return examples def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = 0 for line in test_input_reader: if line.startswith('-DOCSTART-' ) or line == "" or line == "\n": writer.write(__UpperCAmelCase ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: lowerCAmelCase__ :Optional[Any] = line.split()[0] + ' ' + preds_list[example_id].pop(0 ) + '\n' writer.write(__UpperCAmelCase ) else: logger.warning('Maximum sequence length exceeded: No prediction for \'%s\'.' , line.split()[0] ) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if path: with open(__UpperCAmelCase , 'r' ) as f: lowerCAmelCase__ :Any = f.read().splitlines() if "O" not in labels: lowerCAmelCase__ :Union[str, Any] = ['O'] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self ): '''simple docstring''' super().__init__(label_idx=-2 ) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if path: with open(__UpperCAmelCase , 'r' ) as f: lowerCAmelCase__ :str = f.read().splitlines() if "O" not in labels: lowerCAmelCase__ :Optional[Any] = ['O'] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class _lowerCAmelCase ( a ): """simple docstring""" def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if isinstance(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :Union[str, Any] = mode.value lowerCAmelCase__ :Union[str, Any] = os.path.join(__UpperCAmelCase , F"{mode}.txt" ) lowerCAmelCase__ :Any = 1 lowerCAmelCase__ :Optional[Any] = [] with open(__UpperCAmelCase , encoding='utf-8' ) as f: for sentence in parse_incr(__UpperCAmelCase ): lowerCAmelCase__ :Dict = [] lowerCAmelCase__ :Dict = [] for token in sentence: words.append(token['form'] ) labels.append(token['upos'] ) assert len(__UpperCAmelCase ) == len(__UpperCAmelCase ) if words: examples.append(InputExample(guid=F"{mode}-{guid_index}" , words=__UpperCAmelCase , labels=__UpperCAmelCase ) ) guid_index += 1 return examples def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Any = 0 for sentence in parse_incr(__UpperCAmelCase ): lowerCAmelCase__ :Optional[int] = preds_list[example_id] lowerCAmelCase__ :Tuple = '' for token in sentence: out += F"{token['form']} ({token['upos']}|{s_p.pop(0 )}) " out += "\n" writer.write(__UpperCAmelCase ) example_id += 1 def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if path: with open(__UpperCAmelCase , 'r' ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
293
1
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_5_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ] ) class a__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding='''utf-8''' , check=a , ) assert hasattr(self , '''env''' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , a : Tuple ): """simple docstring""" __lowerCamelCase = f"""{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}""" # distributed data settings __lowerCamelCase = {'''smdistributed''': {'''dataparallel''': {'''enabled''': True}}} if self.script != '''run_ddp.py''' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=a , instance_count=a , instance_type=self.instance_type , debugger_hook_config=a , hyperparameters={**self.env.distributed_hyperparameters, '''model_name_or_path''': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=a , py_version='''py36''' , ) def SCREAMING_SNAKE_CASE__ ( self : Dict , a : Any ): """simple docstring""" TrainingJobAnalytics(a ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(2,)] ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , a : Tuple ): """simple docstring""" __lowerCamelCase = self.create_estimator(a ) # run training estimator.fit() # result dataframe __lowerCamelCase = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __lowerCamelCase = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) __lowerCamelCase = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCamelCase = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' , 99_99_99 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , '''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , a )
237
'''simple docstring''' def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> float: if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_to_repay <= 0 or not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise Exception('''Years to repay must be an integer > 0''' ) # Yearly rate is divided by 12 to get monthly rate __lowerCamelCase = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly __lowerCamelCase = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
237
1
"""simple docstring""" import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __UpperCAmelCase ( self ): __a = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_a , '''embed_dim''' ) ) self.parent.assertTrue(hasattr(_a , '''num_heads''' ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self , _a , _a=13 , _a=64 , _a=3 , _a=[16, 48, 96] , _a=[1, 3, 6] , _a=[1, 2, 10] , _a=[7, 3, 3] , _a=[4, 2, 2] , _a=[2, 1, 1] , _a=[2, 2, 2] , _a=[False, False, True] , _a=[0.0, 0.0, 0.0] , _a=0.02 , _a=1E-12 , _a=True , _a=True , _a=2 , ): __a = parent __a = batch_size __a = image_size __a = patch_sizes __a = patch_stride __a = patch_padding __a = is_training __a = use_labels __a = num_labels __a = num_channels __a = embed_dim __a = num_heads __a = stride_kv __a = depth __a = cls_token __a = attention_drop_rate __a = initializer_range __a = layer_norm_eps def __UpperCAmelCase ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.num_labels ) __a = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = CvtModel(config=_a ) model.to(_a ) model.eval() __a = model(_a ) __a = (self.image_size, self.image_size) __a , __a = image_size[0], image_size[1] for i in range(len(self.depth ) ): __a = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __a = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = self.num_labels __a = CvtForImageClassification(_a ) model.to(_a ) model.eval() __a = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCAmelCase ( self ): __a = self.prepare_config_and_inputs() __a , __a , __a = config_and_inputs __a = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = (CvtModel, CvtForImageClassification) if is_torch_available() else () __UpperCAmelCase : List[Any] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : Any = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : List[str] = False def __UpperCAmelCase ( self ): __a = CvtModelTester(self ) __a = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __UpperCAmelCase ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __UpperCAmelCase ( self ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __UpperCAmelCase ( self ): def check_hidden_states_output(_a , _a , _a ): __a = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(_a , _a ) ) __a = outputs.hidden_states __a = len(self.model_tester.depth ) self.assertEqual(len(_a ) , _a ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a = True check_hidden_states_output(_a , _a , _a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __UpperCAmelCase ( self ): pass @slow def __UpperCAmelCase ( self ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = CvtModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def lowercase ( ) -> Optional[Any]: __a = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __UpperCAmelCase ( self ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __UpperCAmelCase ( self ): __a = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_a ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): __a = model(**_a ) # verify the logits __a = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , _a ) __a = torch.tensor([0.9285, 0.9015, -0.3150] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1E-4 ) )
45
"""simple docstring""" from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class __lowerCAmelCase : '''simple docstring''' def __init__( self , _a , ): __a = parent __a = 13 __a = 7 __a = True __a = True __a = True __a = 99 __a = 32 __a = 2 __a = 4 __a = 37 __a = '''gelu''' __a = 0.1 __a = 0.1 __a = 512 __a = 16 __a = 2 __a = 0.02 __a = 3 __a = 4 __a = None def __UpperCAmelCase ( self ): __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length] ) __a = None __a = None __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a = ids_tensor([self.batch_size] , self.num_choices ) __a = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ): ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.prepare_config_and_inputs() __a = True __a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = TFEsmModel(config=_a ) __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __a = model(_a ) __a = [input_ids, input_mask] __a = model(_a ) __a = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a , _a , _a , ): __a = True __a = TFEsmModel(config=_a ) __a = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } __a = model(_a ) __a = [input_ids, input_mask] __a = model(_a , encoder_hidden_states=_a ) # Also check the case where encoder outputs are not passed __a = model(_a , attention_mask=_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = TFEsmForMaskedLM(config=_a ) __a = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = self.num_labels __a = TFEsmForTokenClassification(config=_a ) __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __a = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : int = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) __UpperCAmelCase : Tuple = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : Tuple = False __UpperCAmelCase : Union[str, Any] = False def __UpperCAmelCase ( self ): __a = TFEsmModelTester(self ) __a = ConfigTester(self , config_class=_a , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_a ) @slow def __UpperCAmelCase ( self ): for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = TFEsmModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @unittest.skip('''Protein models do not support embedding resizing.''' ) def __UpperCAmelCase ( self ): pass @unittest.skip('''Protein models do not support embedding resizing.''' ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer __a = model.get_bias() assert isinstance(_a , _a ) for k, v in name.items(): assert isinstance(_a , tf.Variable ) else: __a = model.get_output_embeddings() assert x is None __a = model.get_bias() assert name is None @require_tf class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self ): __a = TFEsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __a = tf.constant([[0, 1, 2, 3, 4, 5]] ) __a = model(_a )[0] __a = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , _a ) # compare the actual values for a slice. __a = tf.constant( [ [ [8.92_1518, -10.58_9814, -6.467_1307], [-6.396_7156, -13.91_1377, -1.121_1915], [-7.78_1247, -13.95_1557, -3.74_0592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def __UpperCAmelCase ( self ): __a = TFEsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __a = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) __a = model(_a )[0] # compare the actual values for a slice. __a = tf.constant( [ [ [0.1444_3092, 0.5412_5327, 0.324_7739], [0.3034_0484, 0.0052_6676, 0.3107_7722], [0.3227_8043, -0.2498_7096, 0.341_4628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
45
1
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class lowerCamelCase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowerCamelCase = 42 _lowerCamelCase = jnp.floataa _lowerCamelCase = True def UpperCamelCase__ ( self ) -> List[Any]: super().setup() A = nn.Dense(5 ,dtype=self.dtype ) def __call__( self ,*lowerCamelCase_ ,**lowerCamelCase_ ) -> int: A = super().__call__(*lowerCamelCase_ ,**lowerCamelCase_ ) A = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class lowerCamelCase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowerCamelCase = FlaxBigBirdForNaturalQuestionsModule def _A ( _a : int , _a : Union[str, Any] , _a : Dict , _a : List[str] , _a : Dict , _a : Tuple ): """simple docstring""" def cross_entropy(_a : Dict , _a : List[str] , _a : Optional[Any]=None ): A = logits.shape[-1] A = (labels[..., None] == jnp.arange(_a )[None]).astype("""f4""" ) A = jax.nn.log_softmax(_a , axis=-1 ) A = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: A = reduction(_a ) return loss A = partial(_a , reduction=jnp.mean ) A = cross_entropy(_a , _a ) A = cross_entropy(_a , _a ) A = cross_entropy(_a , _a ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class lowerCamelCase__ : '''simple docstring''' _lowerCamelCase = '''google/bigbird-roberta-base''' _lowerCamelCase = 3000 _lowerCamelCase = 10500 _lowerCamelCase = 128 _lowerCamelCase = 3 _lowerCamelCase = 1 _lowerCamelCase = 5 # tx_args _lowerCamelCase = 3E-5 _lowerCamelCase = 0.0 _lowerCamelCase = 20000 _lowerCamelCase = 0.0_0_9_5 _lowerCamelCase = '''bigbird-roberta-natural-questions''' _lowerCamelCase = '''training-expt''' _lowerCamelCase = '''data/nq-training.jsonl''' _lowerCamelCase = '''data/nq-validation.jsonl''' def UpperCamelCase__ ( self ) -> int: os.makedirs(self.base_dir ,exist_ok=lowerCamelCase_ ) A = os.path.join(self.base_dir ,self.save_dir ) A = self.batch_size_per_device * jax.device_count() @dataclass class lowerCamelCase__ : '''simple docstring''' _lowerCamelCase = 42 _lowerCamelCase = 4096 # no dynamic padding on TPUs def __call__( self ,lowerCamelCase_ ) -> Tuple: A = self.collate_fn(lowerCamelCase_ ) A = jax.tree_util.tree_map(lowerCamelCase_ ,lowerCamelCase_ ) return batch def UpperCamelCase__ ( self ,lowerCamelCase_ ) -> Dict: A , A = self.fetch_inputs(features["""input_ids"""] ) A = { """input_ids""": jnp.array(lowerCamelCase_ ,dtype=jnp.intaa ), """attention_mask""": jnp.array(lowerCamelCase_ ,dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""] ,dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""] ,dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""] ,dtype=jnp.intaa ), } return batch def UpperCamelCase__ ( self ,lowerCamelCase_ ) -> Union[str, Any]: A = [self._fetch_inputs(lowerCamelCase_ ) for ids in input_ids] return zip(*lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ) -> Tuple: A = [1 for _ in range(len(lowerCamelCase_ ) )] while len(lowerCamelCase_ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def _A ( _a : List[Any] , _a : Tuple , _a : Any=None ): """simple docstring""" if seed is not None: A = dataset.shuffle(seed=_a ) for i in range(len(_a ) // batch_size ): A = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_a ) @partial(jax.pmap , axis_name="""batch""" ) def _A ( _a : Tuple , _a : Any , **_a : Any ): """simple docstring""" def loss_fn(_a : str ): A = model_inputs.pop("""start_labels""" ) A = model_inputs.pop("""end_labels""" ) A = model_inputs.pop("""pooled_labels""" ) A = state.apply_fn(**_a , params=_a , dropout_rng=_a , train=_a ) A , A , A = outputs return state.loss_fn( _a , _a , _a , _a , _a , _a , ) A , A = jax.random.split(_a ) A = jax.value_and_grad(_a ) A , A = grad_fn(state.params ) A = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) A = jax.lax.pmean(_a , """batch""" ) A = state.apply_gradients(grads=_a ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="""batch""" ) def _A ( _a : Dict , **_a : List[Any] ): """simple docstring""" A = model_inputs.pop("""start_labels""" ) A = model_inputs.pop("""end_labels""" ) A = model_inputs.pop("""pooled_labels""" ) A = state.apply_fn(**_a , params=state.params , train=_a ) A , A , A = outputs A = state.loss_fn(_a , _a , _a , _a , _a , _a ) A = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) return metrics class lowerCamelCase__ ( train_state.TrainState ): '''simple docstring''' _lowerCamelCase = struct.field(pytree_node=SCREAMING_SNAKE_CASE ) @dataclass class lowerCamelCase__ : '''simple docstring''' _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = None def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_=None ) -> List[str]: A = model.params A = TrainState.create( apply_fn=model.__call__ ,params=lowerCamelCase_ ,tx=lowerCamelCase_ ,loss_fn=lowerCamelCase_ ,) if ckpt_dir is not None: A , A , A , A , A = restore_checkpoint(lowerCamelCase_ ,lowerCamelCase_ ) A = { """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } A , A = build_tx(**lowerCamelCase_ ) A = train_state.TrainState( step=lowerCamelCase_ ,apply_fn=model.__call__ ,params=lowerCamelCase_ ,tx=lowerCamelCase_ ,opt_state=lowerCamelCase_ ,) A = args A = data_collator A = lr A = params A = jax_utils.replicate(lowerCamelCase_ ) return state def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) -> Any: A = self.args A = len(lowerCamelCase_ ) // args.batch_size A = jax.random.PRNGKey(0 ) A = jax.random.split(lowerCamelCase_ ,jax.device_count() ) for epoch in range(args.max_epochs ): A = jnp.array(0 ,dtype=jnp.floataa ) A = get_batched_dataset(lowerCamelCase_ ,args.batch_size ,seed=lowerCamelCase_ ) A = 0 for batch in tqdm(lowerCamelCase_ ,total=lowerCamelCase_ ,desc=f'Running EPOCH-{epoch}' ): A = self.data_collator(lowerCamelCase_ ) A , A , A = self.train_step_fn(lowerCamelCase_ ,lowerCamelCase_ ,**lowerCamelCase_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: A = jax_utils.unreplicate(state.step ) A = running_loss.item() / i A = self.scheduler_fn(state_step - 1 ) A = self.evaluate(lowerCamelCase_ ,lowerCamelCase_ ) A = { """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(lowerCamelCase_ ) ) self.logger.log(lowerCamelCase_ ,commit=lowerCamelCase_ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' ,state=lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ) -> int: A = get_batched_dataset(lowerCamelCase_ ,self.args.batch_size ) A = len(lowerCamelCase_ ) // self.args.batch_size A = jnp.array(0 ,dtype=jnp.floataa ) A = 0 for batch in tqdm(lowerCamelCase_ ,total=lowerCamelCase_ ,desc="""Evaluating ... """ ): A = self.data_collator(lowerCamelCase_ ) A = self.val_step_fn(lowerCamelCase_ ,**lowerCamelCase_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ) -> List[Any]: A = jax_utils.unreplicate(lowerCamelCase_ ) print(f'SAVING CHECKPOINT IN {save_dir}' ,end=""" ... """ ) self.model_save_fn(lowerCamelCase_ ,params=state.params ) with open(os.path.join(lowerCamelCase_ ,"""opt_state.msgpack""" ) ,"""wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args ,os.path.join(lowerCamelCase_ ,"""args.joblib""" ) ) joblib.dump(self.data_collator ,os.path.join(lowerCamelCase_ ,"""data_collator.joblib""" ) ) with open(os.path.join(lowerCamelCase_ ,"""training_state.json""" ) ,"""w""" ) as f: json.dump({"""step""": state.step.item()} ,lowerCamelCase_ ) print("""DONE""" ) def _A ( _a : Union[str, Any] , _a : Tuple ): """simple docstring""" print(f'RESTORING CHECKPOINT FROM {save_dir}' , end=""" ... """ ) with open(os.path.join(_a , """flax_model.msgpack""" ) , """rb""" ) as f: A = from_bytes(state.params , f.read() ) with open(os.path.join(_a , """opt_state.msgpack""" ) , """rb""" ) as f: A = from_bytes(state.opt_state , f.read() ) A = joblib.load(os.path.join(_a , """args.joblib""" ) ) A = joblib.load(os.path.join(_a , """data_collator.joblib""" ) ) with open(os.path.join(_a , """training_state.json""" ) , """r""" ) as f: A = json.load(_a ) A = training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def _A ( _a : Optional[Any] , _a : Tuple , _a : Tuple , _a : Dict ): """simple docstring""" A = num_train_steps - warmup_steps A = optax.linear_schedule(init_value=_a , end_value=_a , transition_steps=_a ) A = optax.linear_schedule(init_value=_a , end_value=1E-7 , transition_steps=_a ) A = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def _A ( _a : Union[str, Any] , _a : Any , _a : int , _a : List[str] , _a : Optional[int] ): """simple docstring""" def weight_decay_mask(_a : Dict ): A = traverse_util.flatten_dict(_a ) A = {k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(_a ) A = scheduler_fn(_a , _a , _a , _a ) A = optax.adamw(learning_rate=_a , weight_decay=_a , mask=_a ) return tx, lr
361
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL UpperCAmelCase =logging.get_logger(__name__) def _A ( _a : List[str] ): """simple docstring""" if isinstance(_a , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(_a , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(_a ): return [[videos]] raise ValueError(f'Could not make batched video from {videos}' ) class lowerCamelCase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowerCamelCase = ['''pixel_values'''] def __init__( self ,lowerCamelCase_ = True ,lowerCamelCase_ = None ,lowerCamelCase_ = PILImageResampling.BILINEAR ,lowerCamelCase_ = True ,lowerCamelCase_ = None ,lowerCamelCase_ = True ,lowerCamelCase_ = 1 / 2_5_5 ,lowerCamelCase_ = True ,lowerCamelCase_ = True ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,**lowerCamelCase_ ,) -> None: super().__init__(**lowerCamelCase_ ) A = size if size is not None else {"""shortest_edge""": 2_5_6} A = get_size_dict(lowerCamelCase_ ,default_to_square=lowerCamelCase_ ) A = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} A = get_size_dict(lowerCamelCase_ ,param_name="""crop_size""" ) A = do_resize A = size A = do_center_crop A = crop_size A = resample A = do_rescale A = rescale_factor A = offset A = do_normalize A = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN A = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = PILImageResampling.BILINEAR ,lowerCamelCase_ = None ,**lowerCamelCase_ ,) -> np.ndarray: A = get_size_dict(lowerCamelCase_ ,default_to_square=lowerCamelCase_ ) if "shortest_edge" in size: A = get_resize_output_image_size(lowerCamelCase_ ,size["""shortest_edge"""] ,default_to_square=lowerCamelCase_ ) elif "height" in size and "width" in size: A = (size["""height"""], size["""width"""]) else: raise ValueError(f'Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}' ) return resize(lowerCamelCase_ ,size=lowerCamelCase_ ,resample=lowerCamelCase_ ,data_format=lowerCamelCase_ ,**lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = None ,**lowerCamelCase_ ,) -> np.ndarray: A = get_size_dict(lowerCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f'Size must have \'height\' and \'width\' as keys. Got {size.keys()}' ) return center_crop(lowerCamelCase_ ,size=(size["""height"""], size["""width"""]) ,data_format=lowerCamelCase_ ,**lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = True ,lowerCamelCase_ = None ,**lowerCamelCase_ ,) -> List[str]: A = image.astype(np.floataa ) if offset: A = image - (scale / 2) return rescale(lowerCamelCase_ ,scale=lowerCamelCase_ ,data_format=lowerCamelCase_ ,**lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = None ,**lowerCamelCase_ ,) -> np.ndarray: return normalize(lowerCamelCase_ ,mean=lowerCamelCase_ ,std=lowerCamelCase_ ,data_format=lowerCamelCase_ ,**lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = ChannelDimension.FIRST ,) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. A = to_numpy_array(lowerCamelCase_ ) if do_resize: A = self.resize(image=lowerCamelCase_ ,size=lowerCamelCase_ ,resample=lowerCamelCase_ ) if do_center_crop: A = self.center_crop(lowerCamelCase_ ,size=lowerCamelCase_ ) if do_rescale: A = self.rescale(image=lowerCamelCase_ ,scale=lowerCamelCase_ ,offset=lowerCamelCase_ ) if do_normalize: A = self.normalize(image=lowerCamelCase_ ,mean=lowerCamelCase_ ,std=lowerCamelCase_ ) A = to_channel_dimension_format(lowerCamelCase_ ,lowerCamelCase_ ) return image def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = None ,lowerCamelCase_ = ChannelDimension.FIRST ,**lowerCamelCase_ ,) -> PIL.Image.Image: A = do_resize if do_resize is not None else self.do_resize A = resample if resample is not None else self.resample A = do_center_crop if do_center_crop is not None else self.do_center_crop A = do_rescale if do_rescale is not None else self.do_rescale A = rescale_factor if rescale_factor is not None else self.rescale_factor A = offset if offset is not None else self.offset A = do_normalize if do_normalize is not None else self.do_normalize A = image_mean if image_mean is not None else self.image_mean A = image_std if image_std is not None else self.image_std A = size if size is not None else self.size A = get_size_dict(lowerCamelCase_ ,default_to_square=lowerCamelCase_ ) A = crop_size if crop_size is not None else self.crop_size A = get_size_dict(lowerCamelCase_ ,param_name="""crop_size""" ) if not valid_images(lowerCamelCase_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) A = make_batched(lowerCamelCase_ ) A = [ [ self._preprocess_image( image=lowerCamelCase_ ,do_resize=lowerCamelCase_ ,size=lowerCamelCase_ ,resample=lowerCamelCase_ ,do_center_crop=lowerCamelCase_ ,crop_size=lowerCamelCase_ ,do_rescale=lowerCamelCase_ ,rescale_factor=lowerCamelCase_ ,offset=lowerCamelCase_ ,do_normalize=lowerCamelCase_ ,image_mean=lowerCamelCase_ ,image_std=lowerCamelCase_ ,data_format=lowerCamelCase_ ,) for img in video ] for video in videos ] A = {"""pixel_values""": videos} return BatchFeature(data=lowerCamelCase_ ,tensor_type=lowerCamelCase_ )
77
0
from __future__ import annotations import os from typing import Any import requests a : str = 'https://api.github.com' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user a : Optional[Any] = BASE_URL + '/user' # https://github.com/settings/tokens a : Union[str, Any] = os.environ.get('USER_TOKEN', '') def lowerCAmelCase_ (lowerCAmelCase__: str ): """simple docstring""" UpperCAmelCase_: List[Any] = { """Authorization""": F'token {auth_token}', """Accept""": """application/vnd.github.v3+json""", } return requests.get(lowerCAmelCase__ , headers=lowerCAmelCase__ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F'''{key}: {value}''') else: raise ValueError('\'USER_TOKEN\' field cannot be empty.')
147
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Any = logging.get_logger(__name__) class _a ( _lowerCAmelCase ): A = ['''pixel_values'''] def __init__(self, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = 1 / 255, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> None: super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: int = size if size is not None else {"""height""": 256, """width""": 256} UpperCAmelCase_: str = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Union[str, Any] = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} UpperCAmelCase_: Any = get_size_dict(SCREAMING_SNAKE_CASE_, param_name="""crop_size""" ) UpperCAmelCase_: Dict = do_resize UpperCAmelCase_: Tuple = size UpperCAmelCase_: Dict = resample UpperCAmelCase_: Union[str, Any] = do_center_crop UpperCAmelCase_: List[str] = crop_size UpperCAmelCase_: Optional[int] = do_rescale UpperCAmelCase_: Dict = rescale_factor UpperCAmelCase_: Optional[Any] = do_normalize UpperCAmelCase_: Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_: Union[str, Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCAmelCase_: int = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( SCREAMING_SNAKE_CASE_, size=(size["""height"""], size["""width"""]), resample=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCAmelCase_: int = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(SCREAMING_SNAKE_CASE_, size=(size["""height"""], size["""width"""]), data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> str: return rescale(SCREAMING_SNAKE_CASE_, scale=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: return normalize(SCREAMING_SNAKE_CASE_, mean=SCREAMING_SNAKE_CASE_, std=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST, **SCREAMING_SNAKE_CASE_, ) -> PIL.Image.Image: UpperCAmelCase_: str = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_: str = resample if resample is not None else self.resample UpperCAmelCase_: Any = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_: str = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_: List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_: str = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_: Optional[int] = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_: Optional[Any] = image_std if image_std is not None else self.image_std UpperCAmelCase_: List[str] = size if size is not None else self.size UpperCAmelCase_: Any = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[Any] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_: Dict = get_size_dict(SCREAMING_SNAKE_CASE_, param_name="""crop_size""" ) UpperCAmelCase_: Any = make_list_of_images(SCREAMING_SNAKE_CASE_ ) if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase_: List[str] = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] if do_resize: UpperCAmelCase_: Any = [self.resize(image=SCREAMING_SNAKE_CASE_, size=SCREAMING_SNAKE_CASE_, resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_center_crop: UpperCAmelCase_: Dict = [self.center_crop(image=SCREAMING_SNAKE_CASE_, size=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: UpperCAmelCase_: str = [self.rescale(image=SCREAMING_SNAKE_CASE_, scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: UpperCAmelCase_: Optional[int] = [self.normalize(image=SCREAMING_SNAKE_CASE_, mean=SCREAMING_SNAKE_CASE_, std=SCREAMING_SNAKE_CASE_ ) for image in images] UpperCAmelCase_: Dict = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) for image in images] UpperCAmelCase_: Any = {"""pixel_values""": images} return BatchFeature(data=SCREAMING_SNAKE_CASE_, tensor_type=SCREAMING_SNAKE_CASE_ )
147
1
from collections import defaultdict def __lowerCamelCase ( __magic_name__ : Union[str, Any] ): a__: Any =1 a__: str =True for v in tree[start]: if v not in visited: ret += dfs(__UpperCAmelCase ) if ret % 2 == 0: cuts.append(__UpperCAmelCase ) return ret def __lowerCamelCase ( ): dfs(1 ) if __name__ == "__main__": __UpperCAmelCase = 10, 9 __UpperCAmelCase = defaultdict(list) __UpperCAmelCase = {} __UpperCAmelCase = [] __UpperCAmelCase = 0 __UpperCAmelCase = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
355
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''google/mobilenet_v1_1.0_224''': '''https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json''', '''google/mobilenet_v1_0.75_192''': '''https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class lowerCamelCase__ ( _a ): _lowerCAmelCase = '''mobilenet_v1''' def __init__( self : int , _a : Tuple=3 , _a : str=2_2_4 , _a : Dict=1.0 , _a : List[Any]=8 , _a : Tuple="relu6" , _a : Dict=True , _a : Optional[int]=0.9_9_9 , _a : List[Any]=0.0_2 , _a : Optional[Any]=0.0_0_1 , **_a : Optional[int] , ): super().__init__(**_a ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) a__: str =num_channels a__: Union[str, Any] =image_size a__: Dict =depth_multiplier a__: Union[str, Any] =min_depth a__: Any =hidden_act a__: int =tf_padding a__: Dict =classifier_dropout_prob a__: Any =initializer_range a__: List[str] =layer_norm_eps class lowerCamelCase__ ( _a ): _lowerCAmelCase = version.parse('''1.11''' ) @property def _lowerCamelCase ( self : int ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def _lowerCamelCase ( self : Tuple ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def _lowerCamelCase ( self : Dict ): return 1e-4
42
0
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' snake_case_ = VideoMAEConfig() set_architecture_configs(__UpperCAmelCase, __UpperCAmelCase ) if "finetuned" not in model_name: snake_case_ = False if "finetuned" in model_name: snake_case_ = '''huggingface/label-files''' if "kinetics" in model_name: snake_case_ = 400 snake_case_ = '''kinetics400-id2label.json''' elif "ssv2" in model_name: snake_case_ = 174 snake_case_ = '''something-something-v2-id2label.json''' else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''' ) snake_case_ = json.load(open(hf_hub_download(__UpperCAmelCase, __UpperCAmelCase, repo_type='''dataset''' ), '''r''' ) ) snake_case_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> List[str]: '''simple docstring''' if "small" in model_name: snake_case_ = 384 snake_case_ = 1536 snake_case_ = 12 snake_case_ = 16 snake_case_ = 12 snake_case_ = 3 snake_case_ = 192 snake_case_ = 768 elif "large" in model_name: snake_case_ = 1024 snake_case_ = 4096 snake_case_ = 24 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 512 snake_case_ = 2048 elif "huge" in model_name: snake_case_ = 1280 snake_case_ = 5120 snake_case_ = 32 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 640 snake_case_ = 2560 elif "base" not in model_name: raise ValueError('''Model name should include either "small", "base", "large", or "huge"''' ) def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if "encoder." in name: snake_case_ = name.replace('''encoder.''', '''''' ) if "cls_token" in name: snake_case_ = name.replace('''cls_token''', '''videomae.embeddings.cls_token''' ) if "decoder_pos_embed" in name: snake_case_ = name.replace('''decoder_pos_embed''', '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case_ = name.replace('''pos_embed''', '''videomae.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case_ = name.replace('''patch_embed.proj''', '''videomae.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case_ = name.replace('''patch_embed.norm''', '''videomae.embeddings.norm''' ) if "decoder.blocks" in name: snake_case_ = name.replace('''decoder.blocks''', '''decoder.decoder_layers''' ) if "blocks" in name: snake_case_ = name.replace('''blocks''', '''videomae.encoder.layer''' ) if "attn.proj" in name: snake_case_ = name.replace('''attn.proj''', '''attention.output.dense''' ) if "attn" in name and "bias" not in name: snake_case_ = name.replace('''attn''', '''attention.self''' ) if "attn" in name: snake_case_ = name.replace('''attn''', '''attention.attention''' ) if "norm1" in name: snake_case_ = name.replace('''norm1''', '''layernorm_before''' ) if "norm2" in name: snake_case_ = name.replace('''norm2''', '''layernorm_after''' ) if "mlp.fc1" in name: snake_case_ = name.replace('''mlp.fc1''', '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case_ = name.replace('''mlp.fc2''', '''output.dense''' ) if "decoder_embed" in name: snake_case_ = name.replace('''decoder_embed''', '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case_ = name.replace('''decoder_norm''', '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case_ = name.replace('''decoder_pred''', '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.weight''', '''videomae.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.bias''', '''videomae.layernorm.bias''' ) if "head" in name and "decoder" not in name: snake_case_ = name.replace('''head''', '''classifier''' ) return name def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ = orig_state_dict.pop(__UpperCAmelCase ) if key.startswith('''encoder.''' ): snake_case_ = key.replace('''encoder.''', '''''' ) if "qkv" in key: snake_case_ = key.split('''.''' ) if key.startswith('''decoder.blocks''' ): snake_case_ = config.decoder_hidden_size snake_case_ = int(key_split[2] ) snake_case_ = '''decoder.decoder_layers.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = config.hidden_size snake_case_ = int(key_split[1] ) snake_case_ = '''videomae.encoder.layer.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = val return orig_state_dict def __magic_name__ ( ) -> List[Any]: '''simple docstring''' snake_case_ = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''', filename='''eating_spaghetti.npy''', repo_type='''dataset''' ) snake_case_ = np.load(__UpperCAmelCase ) return list(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> Any: '''simple docstring''' snake_case_ = get_videomae_config(__UpperCAmelCase ) if "finetuned" in model_name: snake_case_ = VideoMAEForVideoClassification(__UpperCAmelCase ) else: snake_case_ = VideoMAEForPreTraining(__UpperCAmelCase ) # download original checkpoint, hosted on Google Drive snake_case_ = '''pytorch_model.bin''' gdown.cached_download(__UpperCAmelCase, __UpperCAmelCase, quiet=__UpperCAmelCase ) snake_case_ = torch.load(__UpperCAmelCase, map_location='''cpu''' ) if "model" in files: snake_case_ = files['''model'''] else: snake_case_ = files['''module'''] snake_case_ = convert_state_dict(__UpperCAmelCase, __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) model.eval() # verify model on basic input snake_case_ = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5], image_std=[0.5, 0.5, 0.5] ) snake_case_ = prepare_video() snake_case_ = image_processor(__UpperCAmelCase, return_tensors='''pt''' ) if "finetuned" not in model_name: snake_case_ = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''', filename='''bool_masked_pos.pt''' ) snake_case_ = torch.load(__UpperCAmelCase ) snake_case_ = model(**__UpperCAmelCase ) snake_case_ = outputs.logits snake_case_ = [ '''videomae-small-finetuned-kinetics''', '''videomae-small-finetuned-ssv2''', # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) '''videomae-base-short''', '''videomae-base-short-finetuned-kinetics''', '''videomae-base''', '''videomae-base-finetuned-kinetics''', '''videomae-large''', '''videomae-large-finetuned-kinetics''', '''videomae-huge-finetuned-kinetics''', # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) '''videomae-base-short-ssv2''', '''videomae-base-short-finetuned-ssv2''', '''videomae-base-ssv2''', '''videomae-base-finetuned-ssv2''', ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one snake_case_ = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(F"Model name not supported. Should be one of {model_names}" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3], __UpperCAmelCase, atol=1e-4 ) else: print('''Logits:''', logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3], __UpperCAmelCase, atol=1e-4 ) print('''Logits ok!''' ) # verify loss, if applicable if model_name == "videomae-base-short": snake_case_ = outputs.loss assert torch.allclose(__UpperCAmelCase, __UpperCAmelCase, atol=1e-4 ) print('''Loss ok!''' ) if pytorch_dump_folder_path is not None: print(F"Saving model and image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) if push_to_hub: print('''Pushing to the hub...''' ) model.push_to_hub(__UpperCAmelCase, organization='''nielsr''' ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) a : Union[str, Any] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
56
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class UpperCamelCase_ : """simple docstring""" def __init__( self : Optional[Any] , _lowerCamelCase : List[Any] , _lowerCamelCase : List[str] , _lowerCamelCase : List[str] ): """simple docstring""" if dst_width < 0 or dst_height < 0: raise ValueError('''Destination width/height should be > 0''' ) A_ : int = img A_ : Optional[int] = img.shape[1] A_ : str = img.shape[0] A_ : int = dst_width A_ : Any = dst_height A_ : Tuple = self.src_w / self.dst_w A_ : Any = self.src_h / self.dst_h A_ : Optional[Any] = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 255 ) def _a ( self : Any ): """simple docstring""" for i in range(self.dst_h ): for j in range(self.dst_w ): A_ : List[str] = self.img[self.get_y(a_ )][self.get_x(a_ )] def _a ( self : Dict , _lowerCamelCase : int ): """simple docstring""" return int(self.ratio_x * x ) def _a ( self : Optional[int] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" return int(self.ratio_y * y ) if __name__ == "__main__": snake_case__ = 8_00, 6_00 snake_case__ = imread("""image_data/lena.jpg""", 1) snake_case__ = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F'Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}', n.output ) waitKey(0) destroyAllWindows()
351
'''simple docstring''' import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) snake_case__ = logging.getLogger(__name__) @dataclass(frozen=a__ ) class UpperCamelCase_ : """simple docstring""" _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = None @dataclass(frozen=a__ ) class UpperCamelCase_ : """simple docstring""" _lowerCAmelCase = 42 _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = None if is_torch_available(): import torch from torch.utils.data import Dataset class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = 42 def __init__( self : Optional[int] , _lowerCamelCase : str , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : str , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : List[Any]=False , _lowerCamelCase : bool = False , ): """simple docstring""" A_ : Optional[int] = hans_processors[task]() A_ : int = os.path.join( _lowerCamelCase , '''cached_{}_{}_{}_{}'''.format( '''dev''' if evaluate else '''train''' , tokenizer.__class__.__name__ , str(_lowerCamelCase ) , _lowerCamelCase , ) , ) A_ : Dict = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) A_ ,A_ : List[str] = label_list[2], label_list[1] A_ : Optional[int] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. A_ : str = cached_features_file + '''.lock''' with FileLock(_lowerCamelCase ): if os.path.exists(_lowerCamelCase ) and not overwrite_cache: logger.info(f'Loading features from cached file {cached_features_file}' ) A_ : List[str] = torch.load(_lowerCamelCase ) else: logger.info(f'Creating features from dataset file at {data_dir}' ) A_ : Optional[int] = ( processor.get_dev_examples(_lowerCamelCase ) if evaluate else processor.get_train_examples(_lowerCamelCase ) ) logger.info('''Training examples: %s''' , len(_lowerCamelCase ) ) A_ : Optional[int] = hans_convert_examples_to_features(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) logger.info('''Saving features into cached file %s''' , _lowerCamelCase ) torch.save(self.features , _lowerCamelCase ) def __len__( self : List[str] ): """simple docstring""" return len(self.features ) def __getitem__( self : List[str] , _lowerCamelCase : Optional[int] ): """simple docstring""" return self.features[i] def _a ( self : str ): """simple docstring""" return self.label_list if is_tf_available(): import tensorflow as tf class UpperCamelCase_ : """simple docstring""" _lowerCAmelCase = 42 def __init__( self : Optional[int] , _lowerCamelCase : str , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : str , _lowerCamelCase : Optional[int] = 128 , _lowerCamelCase : Dict=False , _lowerCamelCase : bool = False , ): """simple docstring""" A_ : Optional[int] = hans_processors[task]() A_ : Optional[int] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) A_ ,A_ : Union[str, Any] = label_list[2], label_list[1] A_ : Tuple = label_list A_ : Optional[int] = processor.get_dev_examples(_lowerCamelCase ) if evaluate else processor.get_train_examples(_lowerCamelCase ) A_ : Tuple = hans_convert_examples_to_features(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='''convert examples to features''' ): if ex_index % 10000 == 0: logger.info('''Writing example %d of %d''' % (ex_index, len(_lowerCamelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) A_ : List[Any] = tf.data.Dataset.from_generator( _lowerCamelCase , ( { '''example_id''': tf.intaa, '''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa, }, tf.intaa, ) , ( { '''example_id''': tf.TensorShape([] ), '''input_ids''': tf.TensorShape([None, None] ), '''attention_mask''': tf.TensorShape([None, None] ), '''token_type_ids''': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def _a ( self : Any ): """simple docstring""" return self.dataset def __len__( self : Dict ): """simple docstring""" return len(self.features ) def __getitem__( self : Optional[int] , _lowerCamelCase : List[str] ): """simple docstring""" return self.features[i] def _a ( self : Tuple ): """simple docstring""" return self.label_list class UpperCamelCase_ (a__ ): """simple docstring""" def _a ( self : List[str] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(_lowerCamelCase , '''heuristics_train_set.txt''' ) ) , '''train''' ) def _a ( self : List[str] , _lowerCamelCase : Tuple ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(_lowerCamelCase , '''heuristics_evaluation_set.txt''' ) ) , '''dev''' ) def _a ( self : Any ): """simple docstring""" return ["contradiction", "entailment", "neutral"] def _a ( self : Optional[Any] , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Any ): """simple docstring""" A_ : Tuple = [] for i, line in enumerate(_lowerCamelCase ): if i == 0: continue A_ : str = '''%s-%s''' % (set_type, line[0]) A_ : Optional[Any] = line[5] A_ : Union[str, Any] = line[6] A_ : List[str] = line[7][2:] if line[7].startswith('''ex''' ) else line[7] A_ : str = line[0] examples.append(InputExample(guid=_lowerCamelCase , text_a=_lowerCamelCase , text_b=_lowerCamelCase , label=_lowerCamelCase , pairID=_lowerCamelCase ) ) return examples def snake_case__ ( lowerCamelCase__ : List[InputExample] , lowerCamelCase__ : List[str] , lowerCamelCase__ : int , lowerCamelCase__ : PreTrainedTokenizer , ) -> int: A_ : Union[str, Any] = {label: i for i, label in enumerate(lowerCamelCase__ )} A_ : Optional[Any] = [] for ex_index, example in tqdm.tqdm(enumerate(lowerCamelCase__ ) , desc='''convert examples to features''' ): if ex_index % 1_0_0_0_0 == 0: logger.info('''Writing example %d''' % (ex_index) ) A_ : Optional[int] = tokenizer( example.text_a , example.text_b , add_special_tokens=lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' , truncation=lowerCamelCase__ , return_overflowing_tokens=lowerCamelCase__ , ) A_ : List[str] = label_map[example.label] if example.label in label_map else 0 A_ : Tuple = int(example.pairID ) features.append(InputFeatures(**lowerCamelCase__ , label=lowerCamelCase__ , pairID=lowerCamelCase__ ) ) for i, example in enumerate(examples[:5] ): logger.info('''*** Example ***''' ) logger.info(f'guid: {example}' ) logger.info(f'features: {features[i]}' ) return features snake_case__ = { """hans""": 3, } snake_case__ = { """hans""": HansProcessor, }
4
0
from dataclasses import dataclass, field from typing import Optional @dataclass class _a : _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be trained.'''} ) _lowercase : Optional[str] = field( default='''./''' , metadata={'''help''': '''Save dir where model repo is cloned and models updates are saved to.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot-clean-train''' , metadata={'''help''': '''Name or path of training dataset.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot-clean-valid''' , metadata={'''help''': '''Name or path of validation dataset.'''} ) _lowercase : Optional[int] = field(default=2 , metadata={'''help''': '''Batch size for training.'''} ) _lowercase : Optional[int] = field(default=2 , metadata={'''help''': '''Batch size for evaluation.'''} ) _lowercase : Optional[float] = field(default=0.1 , metadata={'''help''': '''Value of weight decay.'''} ) _lowercase : Optional[int] = field( default=10000 , metadata={'''help''': '''Size of buffer used to shuffle streaming dataset.'''} ) _lowercase : Optional[float] = field(default=2e-4 , metadata={'''help''': '''Learning rate fo training.'''} ) _lowercase : Optional[str] = field(default='''cosine''' , metadata={'''help''': '''Learning rate.'''} ) _lowercase : Optional[int] = field( default=750 , metadata={'''help''': '''Number of warmup steps in the learning rate schedule.'''} ) _lowercase : Optional[int] = field( default=16 , metadata={'''help''': '''Number of gradient accumulation steps.'''} ) _lowercase : Optional[bool] = field( default=UpperCamelCase__ , metadata={'''help''': '''Use gradient checkpointing to reduce memory footprint.'''} ) _lowercase : Optional[int] = field(default=50000 , metadata={'''help''': '''Maximum number of training steps.'''} ) _lowercase : Optional[int] = field( default=-1 , metadata={'''help''': '''Maximum number of evaluation steps. If -1 the full dataset is evaluated.'''} ) _lowercase : Optional[int] = field(default=1024 , metadata={'''help''': '''Sequence lengths used for training.'''} ) _lowercase : Optional[int] = field(default=1 , metadata={'''help''': '''Training seed.'''} ) _lowercase : Optional[int] = field( default=1024 , metadata={'''help''': '''Interval to save checkpoints. Measured as number of forward passes not training steps.'''} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''States path if the training should continue from a checkpoint folder.'''} ) _lowercase : Optional[bool] = field(default=UpperCamelCase__ , metadata={'''help''': '''If True the data is pretokenized.'''} ) @dataclass class _a : _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be evaluated.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot-clean-valid''' , metadata={'''help''': '''Name or path of validation dataset.'''} ) _lowercase : Optional[int] = field(default=2 , metadata={'''help''': '''Batch size used for evaluation.'''} ) _lowercase : Optional[int] = field( default=-1 , metadata={'''help''': '''Maximum number of evaluation steps. If -1 the full dataset is evaluated.'''} ) _lowercase : Optional[int] = field(default=1024 , metadata={'''help''': '''Length of sequences to be evaluated.'''} ) _lowercase : Optional[int] = field(default=1 , metadata={'''help''': '''Random seed used for evaluation.'''} ) @dataclass class _a : _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Model name or path of model to be evaluated.'''} ) _lowercase : Optional[int] = field(default=UpperCamelCase__ , metadata={'''help''': '''Number of workers used for code evaluation.'''} ) _lowercase : Optional[int] = field( default=UpperCamelCase__ , metadata={'''help''': '''The number of human-eval tasks to run. If not included all tasks are evaluated.'''} , ) _lowercase : Optional[bool] = field( default=UpperCamelCase__ , metadata={'''help''': '''Sample from the language model\'s output distribution.'''} ) _lowercase : Optional[float] = field(default=0.2 , metadata={'''help''': '''Sampling temperature used for generation.'''} ) _lowercase : Optional[int] = field(default=256 , metadata={'''help''': '''Maximum number of newly generated tokens.'''} ) _lowercase : Optional[int] = field(default=0 , metadata={'''help''': '''Top-k parameter used for generation.'''} ) _lowercase : Optional[float] = field(default=0.95 , metadata={'''help''': '''Top-p parameter used for nucleus sampling.'''} ) _lowercase : Optional[int] = field(default=10 , metadata={'''help''': '''Number of generations to run in parallel.'''} ) _lowercase : Optional[int] = field( default=200 , metadata={'''help''': '''Number of completions to generate for each sample.'''} ) _lowercase : Optional[int] = field(default=1 , metadata={'''help''': '''Random seed used for evaluation.'''} ) _lowercase : Optional[str] = field( default='''eval_results.json''' , metadata={'''help''': '''Random seed used for evaluation.'''} ) _lowercase : Optional[str] = field( default='''0''' , metadata={'''help''': '''Allow `code_eval` to execute Python code on machine'''} ) _lowercase : Optional[int] = field( default=-1 , metadata={ '''help''': ( '''Determine which device to run the `text-generation` Pipeline on. -1 is CPU and any zero or positive''' ''' number corresponds to which GPU device id to run on.''' ) } , ) @dataclass class _a : _lowercase : Optional[int] = field( default=UpperCamelCase__ , metadata={ '''help''': '''The number of CPU cores to use for parallel preprocessing. Default uses the maximum available.''' } , ) _lowercase : Optional[str] = field( default='''transformersbook/codeparrot''' , metadata={'''help''': '''Folder or name of dataset to process.'''} ) _lowercase : Optional[str] = field( default='''codeparrot-clean''' , metadata={'''help''': '''Folder to save processed processed dataset.'''} ) _lowercase : Optional[int] = field( default=100000 , metadata={'''help''': '''Number of files to save per JSON output file.'''} ) _lowercase : Optional[str] = field(default='''content''' , metadata={'''help''': '''Column containing text data to process.'''} ) _lowercase : Optional[float] = field( default=1000 , metadata={'''help''': '''Maximum line length in file, otherwise file is filtered.'''} ) _lowercase : Optional[float] = field( default=100 , metadata={'''help''': '''Maximum mean line length in file, otherwise file is filtered.'''} ) _lowercase : Optional[float] = field( default=0.25 , metadata={'''help''': '''Maximum fraction of non-alphanumeric characters, otherwise file is filtered.'''} ) _lowercase : Optional[float] = field( default=1.5 , metadata={'''help''': '''Minimum character token ratio for the file, otherwise file is filtered.'''} ) _lowercase : Optional[float] = field( default=0.7 , metadata={'''help''': '''Probability for filtering config, test and uncommon files.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Name or path to the tokenizer.'''} , ) _lowercase : Optional[bool] = field( default=UpperCamelCase__ , metadata={'''help''': '''If True, near-duplicate samples are removed.'''} ) _lowercase : Optional[float] = field( default=0.85 , metadata={'''help''': '''Jaccard threshold for near-duplicate samples.'''} ) @dataclass class _a : _lowercase : Optional[str] = field( default='''gpt2''' , metadata={'''help''': '''Base tokenizer to build new tokenizer from.'''} ) _lowercase : Optional[str] = field( default='''transformersbook/codeparrot-train''' , metadata={'''help''': '''Dataset to train tokenizer on.'''} ) _lowercase : Optional[str] = field(default='''content''' , metadata={'''help''': '''Column containing text data to process.'''} ) _lowercase : Optional[int] = field(default=200000 , metadata={'''help''': '''Number of examples to train tokenizer on.'''} ) _lowercase : Optional[int] = field( default=32768 , metadata={'''help''': '''Number of examples to train the tokenizer on.'''} ) _lowercase : Optional[str] = field(default='''codeparrot''' , metadata={'''help''': '''Name of new tokenizer.'''} ) _lowercase : Optional[bool] = field(default=UpperCamelCase__ , metadata={'''help''': '''Push saved tokenizer to the hub.'''} ) @dataclass class _a : _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Name or path to the tokenizer.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot-clean-train''' , metadata={'''help''': '''Name or path to the dataset to pretokenize.'''} ) _lowercase : Optional[str] = field( default='''tokenized-codeparrot-train''' , metadata={'''help''': '''Repo name of the pretokenized data.'''} ) _lowercase : Optional[int] = field(default=UpperCamelCase__ , metadata={'''help''': '''Number of workers used for code evaluation.'''} ) @dataclass class _a : _lowercase : Optional[str] = field( default='''gpt2-large''' , metadata={'''help''': '''Configuration to use for model initialization.'''} ) _lowercase : Optional[str] = field( default='''codeparrot/codeparrot''' , metadata={'''help''': '''Tokenizer attached to model.'''} ) _lowercase : Optional[str] = field(default='''codeparrot''' , metadata={'''help''': '''Name of the created model.'''} ) _lowercase : Optional[bool] = field(default=UpperCamelCase__ , metadata={'''help''': '''Push saved tokenizer to the hub.'''} )
110
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process lowerCAmelCase = logging.getLogger(__name__) lowerCAmelCase = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) lowerCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _a : _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Leave None if you want to train a model from''' ''' scratch.''' ) } , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(UpperCamelCase__ )} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class _a : _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''The input training data file (a text file).'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={ '''help''': ( '''The input training data files (multiple files in glob format). ''' '''Very often splitting large files to smaller files can prevent tokenizer going out of memory''' ) } , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input evaluation data file to evaluate the perplexity on (a text file).'''} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input train ref data file for whole word mask in Chinese.'''} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input eval ref data file for whole word mask in Chinese.'''} , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Whether distinct lines of text in the dataset are to be handled as distinct sequences.'''} , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Train with masked-language modeling loss instead of language modeling.'''} ) _lowercase : bool = field(default=UpperCamelCase__ , metadata={'''help''': '''Whether ot not to use whole word mask.'''} ) _lowercase : float = field( default=0.15 , metadata={'''help''': '''Ratio of tokens to mask for masked language modeling loss'''} ) _lowercase : float = field( default=1 / 6 , metadata={ '''help''': ( '''Ratio of length of a span of masked tokens to surrounding context length for permutation language''' ''' modeling.''' ) } , ) _lowercase : int = field( default=5 , metadata={'''help''': '''Maximum length of a span of masked tokens for permutation language modeling.'''} ) _lowercase : int = field( default=-1 , metadata={ '''help''': ( '''Optional input sequence length after tokenization.''' '''The training dataset will be truncated in block of this size for training.''' '''Default to the model max input length for single sentence inputs (take into account special tokens).''' ) } , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , ): """simple docstring""" def _dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('''You need to set world whole masking and mlm to True for Chinese Whole Word Mask''' ) return LineByLineWithRefDataset( tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size , ref_path=SCREAMING_SNAKE_CASE , ) return LineByLineTextDataset(tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size ) else: return TextDataset( tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=SCREAMING_SNAKE_CASE , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(SCREAMING_SNAKE_CASE ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def _a ( ): """simple docstring""" lowercase__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( '''Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ''' '''or remove the --do_eval argument.''' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. Use' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , SCREAMING_SNAKE_CASE ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.tokenizer_name: lowercase__ = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another''' ''' script, save it,and load it from here, using --tokenizer_name''' ) if model_args.model_name_or_path: lowercase__ = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) else: logger.info('''Training new model from scratch''' ) lowercase__ = AutoModelWithLMHead.from_config(SCREAMING_SNAKE_CASE ) model.resize_token_embeddings(len(SCREAMING_SNAKE_CASE ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( '''BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the''' '''--mlm flag (masked language modeling).''' ) if data_args.block_size <= 0: lowercase__ = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ = ( get_dataset(SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ = ( get_dataset(SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE , evaluate=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ = DataCollatorForPermutationLanguageModeling( tokenizer=SCREAMING_SNAKE_CASE , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ = DataCollatorForWholeWordMask( tokenizer=SCREAMING_SNAKE_CASE , mlm_probability=data_args.mlm_probability ) else: lowercase__ = DataCollatorForLanguageModeling( tokenizer=SCREAMING_SNAKE_CASE , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ = Trainer( model=SCREAMING_SNAKE_CASE , args=SCREAMING_SNAKE_CASE , data_collator=SCREAMING_SNAKE_CASE , train_dataset=SCREAMING_SNAKE_CASE , eval_dataset=SCREAMING_SNAKE_CASE , prediction_loss_only=SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: lowercase__ = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=SCREAMING_SNAKE_CASE ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase__ = trainer.evaluate() lowercase__ = math.exp(eval_output['''eval_loss'''] ) lowercase__ = {'''perplexity''': perplexity} lowercase__ = os.path.join(training_args.output_dir , '''eval_results_lm.txt''' ) if trainer.is_world_master(): with open(SCREAMING_SNAKE_CASE , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' , SCREAMING_SNAKE_CASE , str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) results.update(SCREAMING_SNAKE_CASE ) return results def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" main() if __name__ == "__main__": main()
110
1
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) ) -> IIRFilter: """simple docstring""" snake_case__ : int = tau * frequency / samplerate snake_case__ : Dict = sin(__snake_case ) snake_case__ : Optional[Any] = cos(__snake_case ) snake_case__ : List[str] = _sin / (2 * q_factor) snake_case__ : Union[str, Any] = (1 - _cos) / 2 snake_case__ : int = 1 - _cos snake_case__ : Optional[int] = 1 + alpha snake_case__ : Union[str, Any] = -2 * _cos snake_case__ : Tuple = 1 - alpha snake_case__ : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) ) -> IIRFilter: """simple docstring""" snake_case__ : Dict = tau * frequency / samplerate snake_case__ : Tuple = sin(__snake_case ) snake_case__ : Dict = cos(__snake_case ) snake_case__ : Union[str, Any] = _sin / (2 * q_factor) snake_case__ : Tuple = (1 + _cos) / 2 snake_case__ : Union[str, Any] = -1 - _cos snake_case__ : List[Any] = 1 + alpha snake_case__ : List[str] = -2 * _cos snake_case__ : Any = 1 - alpha snake_case__ : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) ) -> IIRFilter: """simple docstring""" snake_case__ : str = tau * frequency / samplerate snake_case__ : Any = sin(__snake_case ) snake_case__ : Union[str, Any] = cos(__snake_case ) snake_case__ : Any = _sin / (2 * q_factor) snake_case__ : int = _sin / 2 snake_case__ : Dict = 0 snake_case__ : Optional[int] = -ba snake_case__ : Tuple = 1 + alpha snake_case__ : Union[str, Any] = -2 * _cos snake_case__ : Dict = 1 - alpha snake_case__ : Any = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) ) -> IIRFilter: """simple docstring""" snake_case__ : List[str] = tau * frequency / samplerate snake_case__ : Dict = sin(__snake_case ) snake_case__ : Tuple = cos(__snake_case ) snake_case__ : List[Any] = _sin / (2 * q_factor) snake_case__ : Dict = 1 - alpha snake_case__ : List[str] = -2 * _cos snake_case__ : Dict = 1 + alpha snake_case__ : List[Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) , ) -> IIRFilter: """simple docstring""" snake_case__ : List[str] = tau * frequency / samplerate snake_case__ : int = sin(__snake_case ) snake_case__ : List[str] = cos(__snake_case ) snake_case__ : List[Any] = _sin / (2 * q_factor) snake_case__ : int = 10 ** (gain_db / 40) snake_case__ : str = 1 + alpha * big_a snake_case__ : Tuple = -2 * _cos snake_case__ : Tuple = 1 - alpha * big_a snake_case__ : Tuple = 1 + alpha / big_a snake_case__ : List[Any] = -2 * _cos snake_case__ : Union[str, Any] = 1 - alpha / big_a snake_case__ : Any = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) , ) -> IIRFilter: """simple docstring""" snake_case__ : Union[str, Any] = tau * frequency / samplerate snake_case__ : List[Any] = sin(__snake_case ) snake_case__ : str = cos(__snake_case ) snake_case__ : List[Any] = _sin / (2 * q_factor) snake_case__ : Optional[int] = 10 ** (gain_db / 40) snake_case__ : Any = (big_a + 1) - (big_a - 1) * _cos snake_case__ : List[Any] = (big_a + 1) + (big_a - 1) * _cos snake_case__ : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos snake_case__ : Any = (big_a - 1) + (big_a + 1) * _cos snake_case__ : str = 2 * sqrt(__snake_case ) * alpha snake_case__ : Tuple = big_a * (pmc + aaa) snake_case__ : List[Any] = 2 * big_a * mpc snake_case__ : Dict = big_a * (pmc - aaa) snake_case__ : Optional[Any] = ppmc + aaa snake_case__ : str = -2 * pmpc snake_case__ : Any = ppmc - aaa snake_case__ : Optional[int] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 1 / sqrt(2 ) , ) -> IIRFilter: """simple docstring""" snake_case__ : Union[str, Any] = tau * frequency / samplerate snake_case__ : Dict = sin(__snake_case ) snake_case__ : Union[str, Any] = cos(__snake_case ) snake_case__ : List[Any] = _sin / (2 * q_factor) snake_case__ : str = 10 ** (gain_db / 40) snake_case__ : Any = (big_a + 1) - (big_a - 1) * _cos snake_case__ : List[Any] = (big_a + 1) + (big_a - 1) * _cos snake_case__ : Union[str, Any] = (big_a - 1) - (big_a + 1) * _cos snake_case__ : Tuple = (big_a - 1) + (big_a + 1) * _cos snake_case__ : List[Any] = 2 * sqrt(__snake_case ) * alpha snake_case__ : List[Any] = big_a * (ppmc + aaa) snake_case__ : List[str] = -2 * big_a * pmpc snake_case__ : int = big_a * (ppmc - aaa) snake_case__ : List[str] = pmc + aaa snake_case__ : Any = 2 * mpc snake_case__ : List[str] = pmc - aaa snake_case__ : Optional[int] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
356
A__ = 0 # The first color of the flag. A__ = 1 # The second color of the flag. A__ = 2 # The third color of the flag. A__ = (red, white, blue) def _lowerCAmelCase ( __lowerCAmelCase ) -> list: """simple docstring""" if not sequence: return [] if len(__lowerCAmelCase ) == 1: return list(__lowerCAmelCase ) snake_case__ : List[Any] = 0 snake_case__ : str = len(__lowerCAmelCase ) - 1 snake_case__ : List[Any] = 0 while mid <= high: if sequence[mid] == colors[0]: snake_case__ , snake_case__ : List[Any] = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: snake_case__ , snake_case__ : int = sequence[high], sequence[mid] high -= 1 else: snake_case__ : List[Any] = f"""The elements inside the sequence must contains only {colors} values""" raise ValueError(__lowerCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod() A__ = input('''Enter numbers separated by commas:\n''').strip() A__ = [int(item.strip()) for item in user_input.split(''',''')] print(f"""{dutch_national_flag_sort(unsorted)}""")
44
0
'''simple docstring''' import argparse import ast import logging import os import sys import pandas as pd import torch from tqdm import tqdm from transformers import BartForConditionalGeneration, RagRetriever, RagSequenceForGeneration, RagTokenForGeneration from transformers import logging as transformers_logging sys.path.append(os.path.join(os.getcwd())) # noqa: E402 # isort:skip from utils_rag import exact_match_score, fa_score # noqa: E402 # isort:skip __lowercase : Dict = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) transformers_logging.set_verbosity_info() def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] ): if "token" in model_name_or_path: return "rag_token" if "sequence" in model_name_or_path: return "rag_sequence" if "bart" in model_name_or_path: return "bart" return None def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : List[Any] ): return max(metric_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for gt in ground_truths ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[str] ): __a : List[Any] = [line.strip() for line in open(_SCREAMING_SNAKE_CASE , 'r' ).readlines()] __a : List[Any] = [] if args.gold_data_mode == "qa": __a : Any = pd.read_csv(_SCREAMING_SNAKE_CASE , sep='\t' , header=_SCREAMING_SNAKE_CASE ) for answer_list in data[1]: __a : Union[str, Any] = ast.literal_eval(_SCREAMING_SNAKE_CASE ) answers.append(_SCREAMING_SNAKE_CASE ) else: __a : Optional[int] = [line.strip() for line in open(_SCREAMING_SNAKE_CASE , 'r' ).readlines()] __a : Optional[int] = [[reference] for reference in references] __a : List[Any] = 0 for prediction, ground_truths in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): total += 1 em += metric_max_over_ground_truths(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) fa += metric_max_over_ground_truths(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __a : Tuple = 1_0_0.0 * em / total __a : Any = 1_0_0.0 * fa / total logger.info(F"""F1: {fa:.2f}""" ) logger.info(F"""EM: {em:.2f}""" ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Optional[Any] ): __a : str = args.k __a : Union[str, Any] = [line.strip() for line in open(_SCREAMING_SNAKE_CASE , 'r' ).readlines()] __a : Tuple = [line.strip() for line in open(_SCREAMING_SNAKE_CASE , 'r' ).readlines()] __a : Optional[int] = 0 for hypo, reference in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __a : Optional[Any] = set(hypo.split('\t' )[:k] ) __a : Union[str, Any] = set(reference.split('\t' ) ) total += 1 em += len(hypo_provenance & ref_provenance ) / k __a : Tuple = 1_0_0.0 * em / total logger.info(F"""Precision@{k}: {em: .2f}""" ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Optional[int] ): def strip_title(_SCREAMING_SNAKE_CASE : int ): if title.startswith('"' ): __a : int = title[1:] if title.endswith('"' ): __a : Union[str, Any] = title[:-1] return title __a : str = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( _SCREAMING_SNAKE_CASE , return_tensors='pt' , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , )['input_ids'].to(args.device ) __a : Tuple = rag_model.rag.question_encoder(_SCREAMING_SNAKE_CASE ) __a : Tuple = question_enc_outputs[0] __a : int = rag_model.retriever( _SCREAMING_SNAKE_CASE , question_enc_pool_output.cpu().detach().to(torch.floataa ).numpy() , prefix=rag_model.rag.generator.config.prefix , n_docs=rag_model.config.n_docs , return_tensors='pt' , ) __a : Tuple = rag_model.retriever.index.get_doc_dicts(result.doc_ids ) __a : Tuple = [] for docs in all_docs: __a : Dict = [strip_title(_SCREAMING_SNAKE_CASE ) for title in docs['title']] provenance_strings.append('\t'.join(_SCREAMING_SNAKE_CASE ) ) return provenance_strings def lowerCamelCase (_SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Any ): with torch.no_grad(): __a : Dict = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( _SCREAMING_SNAKE_CASE , return_tensors='pt' , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE ) __a : Union[str, Any] = inputs_dict.input_ids.to(args.device ) __a : str = inputs_dict.attention_mask.to(args.device ) __a : Dict = rag_model.generate( # rag_model overwrites generate _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , num_beams=args.num_beams , min_length=args.min_length , max_length=args.max_length , early_stopping=_SCREAMING_SNAKE_CASE , num_return_sequences=1 , bad_words_ids=[[0, 0]] , ) __a : Optional[Any] = rag_model.retriever.generator_tokenizer.batch_decode(_SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE ) if args.print_predictions: for q, a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info('Q: {} - A: {}'.format(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) return answers def lowerCamelCase (): __a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '--model_type' , choices=['rag_sequence', 'rag_token', 'bart'] , type=_SCREAMING_SNAKE_CASE , help=( 'RAG model type: rag_sequence, rag_token or bart, if none specified, the type is inferred from the' ' model_name_or_path' ) , ) parser.add_argument( '--index_name' , default=_SCREAMING_SNAKE_CASE , choices=['exact', 'compressed', 'legacy'] , type=_SCREAMING_SNAKE_CASE , help='RAG model retriever type' , ) parser.add_argument( '--index_path' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , help='Path to the retrieval index' , ) parser.add_argument('--n_docs' , default=5 , type=_SCREAMING_SNAKE_CASE , help='Number of retrieved docs' ) parser.add_argument( '--model_name_or_path' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='Path to pretrained checkpoints or model identifier from huggingface.co/models' , ) parser.add_argument( '--eval_mode' , choices=['e2e', 'retrieval'] , default='e2e' , type=_SCREAMING_SNAKE_CASE , help=( 'Evaluation mode, e2e calculates exact match and F1 of the downstream task, retrieval calculates' ' precision@k.' ) , ) parser.add_argument('--k' , default=1 , type=_SCREAMING_SNAKE_CASE , help='k for the precision@k calculation' ) parser.add_argument( '--evaluation_set' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='Path to a file containing evaluation samples' , ) parser.add_argument( '--gold_data_path' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='Path to a tab-separated file with gold samples' , ) parser.add_argument( '--gold_data_mode' , default='qa' , type=_SCREAMING_SNAKE_CASE , choices=['qa', 'ans'] , help=( 'Format of the gold data file' 'qa - a single line in the following format: question [tab] answer_list' 'ans - a single line of the gold file contains the expected answer string' ) , ) parser.add_argument( '--predictions_path' , type=_SCREAMING_SNAKE_CASE , default='predictions.txt' , help='Name of the predictions file, to be stored in the checkpoints directory' , ) parser.add_argument( '--eval_all_checkpoints' , action='store_true' , help='Evaluate all checkpoints starting with the same prefix as model_name ending and ending with step number' , ) parser.add_argument( '--eval_batch_size' , default=8 , type=_SCREAMING_SNAKE_CASE , help='Batch size per GPU/CPU for evaluation.' , ) parser.add_argument( '--recalculate' , help='Recalculate predictions even if the prediction file exists' , action='store_true' , ) parser.add_argument( '--num_beams' , default=4 , type=_SCREAMING_SNAKE_CASE , help='Number of beams to be used when generating answers' , ) parser.add_argument('--min_length' , default=1 , type=_SCREAMING_SNAKE_CASE , help='Min length of the generated answers' ) parser.add_argument('--max_length' , default=50 , type=_SCREAMING_SNAKE_CASE , help='Max length of the generated answers' ) parser.add_argument( '--print_predictions' , action='store_true' , help='If True, prints predictions while evaluating.' , ) parser.add_argument( '--print_docs' , action='store_true' , help='If True, prints docs retried while generating.' , ) __a : Dict = parser.parse_args() __a : Dict = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) return args def lowerCamelCase (_SCREAMING_SNAKE_CASE : Union[str, Any] ): __a : Tuple = {} if args.model_type is None: __a : str = infer_model_type(args.model_name_or_path ) assert args.model_type is not None if args.model_type.startswith('rag' ): __a : int = RagTokenForGeneration if args.model_type == 'rag_token' else RagSequenceForGeneration __a : int = args.n_docs if args.index_name is not None: __a : Optional[Any] = args.index_name if args.index_path is not None: __a : List[str] = args.index_path else: __a : Tuple = BartForConditionalGeneration __a : int = ( [f.path for f in os.scandir(args.model_name_or_path ) if f.is_dir()] if args.eval_all_checkpoints else [args.model_name_or_path] ) logger.info('Evaluate the following checkpoints: %s' , _SCREAMING_SNAKE_CASE ) __a : str = get_scores if args.eval_mode == 'e2e' else get_precision_at_k __a : Optional[int] = evaluate_batch_eae if args.eval_mode == 'e2e' else evaluate_batch_retrieval for checkpoint in checkpoints: if os.path.exists(args.predictions_path ) and (not args.recalculate): logger.info('Calculating metrics based on an existing predictions file: {}'.format(args.predictions_path ) ) score_fn(_SCREAMING_SNAKE_CASE , args.predictions_path , args.gold_data_path ) continue logger.info('***** Running evaluation for {} *****'.format(_SCREAMING_SNAKE_CASE ) ) logger.info(' Batch size = %d' , args.eval_batch_size ) logger.info(' Predictions will be stored under {}'.format(args.predictions_path ) ) if args.model_type.startswith('rag' ): __a : Tuple = RagRetriever.from_pretrained(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) __a : Union[str, Any] = model_class.from_pretrained(_SCREAMING_SNAKE_CASE , retriever=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) model.retriever.init_retrieval() else: __a : Union[str, Any] = model_class.from_pretrained(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) model.to(args.device ) with open(args.evaluation_set , 'r' ) as eval_file, open(args.predictions_path , 'w' ) as preds_file: __a : Tuple = [] for line in tqdm(_SCREAMING_SNAKE_CASE ): questions.append(line.strip() ) if len(_SCREAMING_SNAKE_CASE ) == args.eval_batch_size: __a : List[Any] = evaluate_batch_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) preds_file.write('\n'.join(_SCREAMING_SNAKE_CASE ) + '\n' ) preds_file.flush() __a : Dict = [] if len(_SCREAMING_SNAKE_CASE ) > 0: __a : Dict = evaluate_batch_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) preds_file.write('\n'.join(_SCREAMING_SNAKE_CASE ) ) preds_file.flush() score_fn(_SCREAMING_SNAKE_CASE , args.predictions_path , args.gold_data_path ) if __name__ == "__main__": __lowercase : Optional[Any] = get_args() main(args)
27
"""simple docstring""" class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [ [], [], [], ] def snake_case ( self , __a , __a ): try: if len(self.queues[priority] ) >= 1_00: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(__a ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def snake_case ( self ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self ): return "\n".join(f"Priority {i}: {q}" for i, q in enumerate(self.queues ) ) class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [] def snake_case ( self , __a ): if len(self.queue ) == 1_00: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(__a ) def snake_case ( self ): if not self.queue: raise UnderFlowError("The queue is empty" ) else: __lowerCAmelCase = min(self.queue ) self.queue.remove(__a ) return data def __str__( self ): return str(self.queue ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
57
0
'''simple docstring''' import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate __A : Any = trt.Logger(trt.Logger.WARNING) __A : List[Any] = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) __A : str = logging.getLogger(__name__) __A : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--onnx_model_path', default=None, type=str, required=True, help='Path to ONNX model: ', ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='The output directory where the model checkpoints and predictions will be written.', ) # Other parameters parser.add_argument( '--tokenizer_name', default='', type=str, required=True, help='Pretrained tokenizer name or path if not the same as model_name', ) parser.add_argument( '--version_2_with_negative', action='store_true', help='If true, the SQuAD examples contain some that do not have an answer.', ) parser.add_argument( '--null_score_diff_threshold', type=float, default=0.0, help='If null_score - best_non_null is greater than the threshold predict null.', ) parser.add_argument( '--max_seq_length', default=384, type=int, help=( 'The maximum total input sequence length after WordPiece tokenization. Sequences ' 'longer than this will be truncated, and sequences shorter than this will be padded.' ), ) parser.add_argument( '--doc_stride', default=128, type=int, help='When splitting up a long document into chunks, how much stride to take between chunks.', ) parser.add_argument('--per_device_eval_batch_size', default=8, type=int, help='Batch size per GPU/CPU for evaluation.') parser.add_argument( '--n_best_size', default=20, type=int, help='The total number of n-best predictions to generate in the nbest_predictions.json output file.', ) parser.add_argument( '--max_answer_length', default=30, type=int, help=( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ), ) parser.add_argument('--seed', type=int, default=42, help='random seed for initialization') parser.add_argument( '--dataset_name', type=str, default=None, required=True, help='The name of the dataset to use (via the datasets library).', ) parser.add_argument( '--dataset_config_name', type=str, default=None, help='The configuration name of the dataset to use (via the datasets library).', ) parser.add_argument( '--preprocessing_num_workers', type=int, default=4, help='A csv or a json file containing the training data.' ) parser.add_argument('--overwrite_cache', action='store_true', help='Overwrite the cached training and evaluation sets') parser.add_argument( '--fp16', action='store_true', help='Whether to use 16-bit (mixed) precision instead of 32-bit', ) parser.add_argument( '--int8', action='store_true', help='Whether to use INT8', ) __A : List[str] = parser.parse_args() if args.tokenizer_name: __A : List[str] = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported by this script.' 'You can do it from another script, save it, and load it from here, using --tokenizer_name.' ) logger.info('Training/evaluation parameters %s', args) __A : Union[str, Any] = args.per_device_eval_batch_size __A : Any = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties __A : List[Any] = True __A : Dict = 'temp_engine/bert-fp32.engine' if args.fpaa: __A : Any = 'temp_engine/bert-fp16.engine' if args.inta: __A : Union[str, Any] = 'temp_engine/bert-int8.engine' # import ONNX file if not os.path.exists('temp_engine'): os.makedirs('temp_engine') __A : List[str] = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, 'rb') as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network __A : int = [network.get_input(i) for i in range(network.num_inputs)] __A : Optional[Any] = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: __A : Tuple = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) __A : List[Any] = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) __A : str = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, 'wb') as f: f.write(engine.serialize()) def UpperCAmelCase ( lowerCamelCase_ :Tuple , lowerCamelCase_ :List[str] , lowerCamelCase_ :List[Any] , lowerCamelCase_ :str , lowerCamelCase_ :Optional[int] , lowerCamelCase_ :str , lowerCamelCase_ :List[str] , lowerCamelCase_ :List[str] ): '''simple docstring''' snake_case_ : Tuple = np.asarray(inputs["""input_ids"""] , dtype=np.intaa ) snake_case_ : List[str] = np.asarray(inputs["""attention_mask"""] , dtype=np.intaa ) snake_case_ : Optional[Any] = np.asarray(inputs["""token_type_ids"""] , dtype=np.intaa ) # Copy inputs cuda.memcpy_htod_async(d_inputs[0] , input_ids.ravel() , lowerCamelCase_ ) cuda.memcpy_htod_async(d_inputs[1] , attention_mask.ravel() , lowerCamelCase_ ) cuda.memcpy_htod_async(d_inputs[2] , token_type_ids.ravel() , lowerCamelCase_ ) # start time snake_case_ : int = time.time() # Run inference context.execute_async( bindings=[int(lowerCamelCase_ ) for d_inp in d_inputs] + [int(lowerCamelCase_ ), int(lowerCamelCase_ )] , stream_handle=stream.handle ) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) cuda.memcpy_dtoh_async(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Synchronize the stream and take time stream.synchronize() # end time snake_case_ : Dict = time.time() snake_case_ : List[Any] = end_time - start_time snake_case_ : Tuple = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. __A : int = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. __A : Dict = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError('Evaluation requires a dataset name') # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. __A : str = raw_datasets['validation'].column_names __A : Optional[Any] = 'question' if 'question' in column_names else column_names[0] __A : Any = 'context' if 'context' in column_names else column_names[1] __A : str = 'answers' if 'answers' in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). __A : Optional[Any] = tokenizer.padding_side == 'right' if args.max_seq_length > tokenizer.model_max_length: logger.warning( F'The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the' F'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) __A : List[Any] = min(args.max_seq_length, tokenizer.model_max_length) def UpperCAmelCase ( lowerCamelCase_ :Tuple ): '''simple docstring''' snake_case_ : Union[str, Any] = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. snake_case_ : List[str] = tokenizer( examples[question_column_name if pad_on_right else context_column_name] , examples[context_column_name if pad_on_right else question_column_name] , truncation="""only_second""" if pad_on_right else """only_first""" , max_length=lowerCamelCase_ , stride=args.doc_stride , return_overflowing_tokens=lowerCamelCase_ , return_offsets_mapping=lowerCamelCase_ , padding="""max_length""" , ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. snake_case_ : List[str] = tokenized_examples.pop("""overflow_to_sample_mapping""" ) # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. snake_case_ : Dict = [] for i in range(len(tokenized_examples["""input_ids"""] ) ): # Grab the sequence corresponding to that example (to know what is the context and what is the question). snake_case_ : int = tokenized_examples.sequence_ids(lowerCamelCase_ ) snake_case_ : Any = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. snake_case_ : Union[str, Any] = sample_mapping[i] tokenized_examples["example_id"].append(examples["""id"""][sample_index] ) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. snake_case_ : Optional[int] = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples["""offset_mapping"""][i] ) ] return tokenized_examples __A : List[Any] = raw_datasets['validation'] # Validation Feature Creation __A : Tuple = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc='Running tokenizer on validation dataset', ) __A : Dict = default_data_collator __A : List[str] = eval_dataset.remove_columns(['example_id', 'offset_mapping']) __A : Optional[int] = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def UpperCAmelCase ( lowerCamelCase_ :Tuple , lowerCamelCase_ :Any , lowerCamelCase_ :Dict , lowerCamelCase_ :Any="eval" ): '''simple docstring''' snake_case_ : int = postprocess_qa_predictions( examples=lowerCamelCase_ , features=lowerCamelCase_ , predictions=lowerCamelCase_ , version_2_with_negative=args.version_2_with_negative , n_best_size=args.n_best_size , max_answer_length=args.max_answer_length , null_score_diff_threshold=args.null_score_diff_threshold , output_dir=args.output_dir , prefix=lowerCamelCase_ , ) # Format the result to the format the metric expects. if args.version_2_with_negative: snake_case_ : Any = [ {"""id""": k, """prediction_text""": v, """no_answer_probability""": 0.0} for k, v in predictions.items() ] else: snake_case_ : Union[str, Any] = [{"""id""": k, """prediction_text""": v} for k, v in predictions.items()] snake_case_ : Any = [{"""id""": ex["""id"""], """answers""": ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=lowerCamelCase_ , label_ids=lowerCamelCase_ ) __A : Any = load_metric('squad_v2' if args.version_2_with_negative else 'squad') # Evaluation! logger.info('Loading ONNX model %s for evaluation', args.onnx_model_path) with open(engine_name, 'rb') as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def UpperCAmelCase ( lowerCamelCase_ :Optional[Any] ): '''simple docstring''' return trt.volume(engine.get_binding_shape(lowerCamelCase_ ) ) * engine.get_binding_dtype(lowerCamelCase_ ).itemsize # Allocate device memory for inputs and outputs. __A : Optional[Any] = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer __A : Optional[int] = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) __A : Union[str, Any] = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) __A : Optional[int] = cuda.mem_alloc(h_outputa.nbytes) __A : Union[str, Any] = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. __A : Any = cuda.Stream() # Evaluation logger.info('***** Running Evaluation *****') logger.info(F' Num examples = {len(eval_dataset)}') logger.info(F' Batch size = {args.per_device_eval_batch_size}') __A : Union[str, Any] = 0.0 __A : Tuple = 0 __A : str = timeit.default_timer() __A : Any = None for step, batch in enumerate(eval_dataloader): __A : Any = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 __A : Dict = outputs __A : Union[str, Any] = torch.tensor(start_logits) __A : Optional[Any] = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered __A : Any = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-100) __A : List[Any] = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-100) __A : Optional[int] = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) __A : List[Any] = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-100) if all_preds is not None: __A : str = nested_truncate(all_preds, len(eval_dataset)) __A : int = timeit.default_timer() - start_time logger.info(' Evaluation done in total %f secs (%f sec per example)', evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info('Average Inference Time = {:.3f} ms'.format(total_time * 1_000 / niter)) logger.info('Total Inference Time = {:.3f} ms'.format(total_time * 1_000)) logger.info('Total Number of Inference = %d', niter) __A : Any = post_processing_function(eval_examples, eval_dataset, all_preds) __A : Union[str, Any] = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(F'Evaluation metrics: {eval_metric}')
354
'''simple docstring''' from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig __A : Dict = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorch': 'https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json', } class __UpperCamelCase ( lowercase__ ): lowercase : Optional[int] = 'ernie_m' lowercase : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self :Optional[Any] ,_UpperCamelCase :int = 2_5_0_0_0_2 ,_UpperCamelCase :int = 7_6_8 ,_UpperCamelCase :int = 1_2 ,_UpperCamelCase :int = 1_2 ,_UpperCamelCase :int = 3_0_7_2 ,_UpperCamelCase :str = "gelu" ,_UpperCamelCase :float = 0.1 ,_UpperCamelCase :float = 0.1 ,_UpperCamelCase :int = 5_1_4 ,_UpperCamelCase :float = 0.02 ,_UpperCamelCase :int = 1 ,_UpperCamelCase :float = 1E-0_5 ,_UpperCamelCase :List[Any]=None ,_UpperCamelCase :List[str]=False ,_UpperCamelCase :Optional[int]=0.0 ,**_UpperCamelCase :List[Any] ,): super().__init__(pad_token_id=_UpperCamelCase ,**_UpperCamelCase ) snake_case_ : Optional[int] = vocab_size snake_case_ : Any = hidden_size snake_case_ : Union[str, Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Any = intermediate_size snake_case_ : Any = hidden_act snake_case_ : Tuple = hidden_dropout_prob snake_case_ : Union[str, Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : int = initializer_range snake_case_ : Optional[Any] = layer_norm_eps snake_case_ : Union[str, Any] = classifier_dropout snake_case_ : Tuple = is_decoder snake_case_ : int = act_dropout
8
0
"""simple docstring""" import unittest import numpy as np def snake_case_ ( A_ : np.ndarray, A_ : np.ndarray, A_ : np.ndarray, A_ : np.ndarray | None = None, ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = np.shape(A_ ) _lowerCamelCase : List[str] = np.shape(A_ ) _lowerCamelCase : List[str] = np.shape(A_ ) if shape_a[0] != shape_b[0]: _lowerCamelCase : Tuple = ( '''Expected the same number of rows for A and B. ''' F'''Instead found A of size {shape_a} and B of size {shape_b}''' ) raise ValueError(A_ ) if shape_b[1] != shape_c[1]: _lowerCamelCase : Tuple = ( '''Expected the same number of columns for B and C. ''' F'''Instead found B of size {shape_b} and C of size {shape_c}''' ) raise ValueError(A_ ) _lowerCamelCase : List[str] = pseudo_inv if a_inv is None: try: _lowerCamelCase : Any = np.linalg.inv(A_ ) except np.linalg.LinAlgError: raise ValueError( '''Input matrix A is not invertible. Cannot compute Schur complement.''' ) return mat_c - mat_b.T @ a_inv @ mat_b class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : List[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _lowerCamelCase : List[str] = np.array([[0, 3], [3, 0], [2, 3]] ) _lowerCamelCase : List[str] = np.array([[2, 1], [6, 3]] ) _lowerCamelCase : List[Any] = schur_complement(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : Dict = np.block([[a, b], [b.T, c]] ) _lowerCamelCase : Tuple = np.linalg.det(__lowerCAmelCase ) _lowerCamelCase : List[str] = np.linalg.det(__lowerCAmelCase ) _lowerCamelCase : Any = np.linalg.det(__lowerCAmelCase ) self.assertAlmostEqual(__lowerCAmelCase , det_a * det_s ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : List[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _lowerCamelCase : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _lowerCamelCase : int = np.array([[2, 1], [6, 3]] ) with self.assertRaises(__lowerCAmelCase ): schur_complement(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _lowerCamelCase : List[str] = np.array([[0, 3], [3, 0], [2, 3]] ) _lowerCamelCase : Union[str, Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(__lowerCAmelCase ): schur_complement(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
72
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ = { '''configuration_autoformer''': [ '''AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AutoformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AutoformerForPrediction''', '''AutoformerModel''', '''AutoformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
"""simple docstring""" from math import log from scipy.constants import Boltzmann, physical_constants A_ : Tuple =3_0_0 # TEMPERATURE (unit = K) def SCREAMING_SNAKE_CASE_ ( snake_case : float , snake_case : float , snake_case : float , )-> float: if donor_conc <= 0: raise ValueError('Donor concentration should be positive' ) elif acceptor_conc <= 0: raise ValueError('Acceptor concentration should be positive' ) elif intrinsic_conc <= 0: raise ValueError('Intrinsic concentration should be positive' ) elif donor_conc <= intrinsic_conc: raise ValueError( 'Donor concentration should be greater than intrinsic concentration' ) elif acceptor_conc <= intrinsic_conc: raise ValueError( 'Acceptor concentration should be greater than intrinsic concentration' ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
80
"""simple docstring""" import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor A_ : Union[str, Any] =logging.get_logger(__name__) class __a ( lowerCAmelCase__ ): def __init__( self , *a__ , **a__ ): warnings.warn( 'The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DPTImageProcessor instead.' , a__ , ) super().__init__(*a__ , **a__ )
80
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ : Dict = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : Dict = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys snake_case__ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
117
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ :str = { '''configuration_megatron_bert''': ['''MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegatronBertConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :Union[str, Any] = [ '''MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegatronBertForCausalLM''', '''MegatronBertForMaskedLM''', '''MegatronBertForMultipleChoice''', '''MegatronBertForNextSentencePrediction''', '''MegatronBertForPreTraining''', '''MegatronBertForQuestionAnswering''', '''MegatronBertForSequenceClassification''', '''MegatronBertForTokenClassification''', '''MegatronBertModel''', '''MegatronBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys lowerCAmelCase__ :List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
329
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__a ) class lowercase_ ( __a ): """simple docstring""" lowerCamelCase_ = field(default='''automatic-speech-recognition''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowerCamelCase_ = Features({'''audio''': Audio()} ) lowerCamelCase_ = Features({'''transcription''': Value('''string''' )} ) lowerCamelCase_ = '''audio''' lowerCamelCase_ = '''transcription''' def lowerCAmelCase_ ( self : Dict , __lowerCamelCase : Optional[Any] ): """simple docstring""" if self.audio_column not in features: raise ValueError(F"""Column {self.audio_column} is not present in features.""" ) if not isinstance(features[self.audio_column] , UpperCamelCase__ ): raise ValueError(F"""Column {self.audio_column} is not an Audio type.""" ) _SCREAMING_SNAKE_CASE = copy.deepcopy(self ) _SCREAMING_SNAKE_CASE = self.input_schema.copy() _SCREAMING_SNAKE_CASE = features[self.audio_column] _SCREAMING_SNAKE_CASE = input_schema return task_template @property def lowerCAmelCase_ ( self : str ): """simple docstring""" return {self.audio_column: "audio", self.transcription_column: "transcription"}
369
'''simple docstring''' import random def SCREAMING_SNAKE_CASE_ ( __A : int , __A : float , __A : bool = False ) -> dict: _SCREAMING_SNAKE_CASE = {i: [] for i in range(__A )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(__A ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(__A ): for j in range(i + 1 , __A ): if random.random() < probability: graph[i].append(__A ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(__A ) return graph def SCREAMING_SNAKE_CASE_ ( __A : int ) -> dict: return { i: [j for j in range(__A ) if i != j] for i in range(__A ) } if __name__ == "__main__": import doctest doctest.testmod()
111
0
"""simple docstring""" from math import isqrt, loga def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' lowerCAmelCase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): lowerCAmelCase = False return [i for i in range(2 , SCREAMING_SNAKE_CASE ) if is_prime[i]] def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : int = 80_08_00 , SCREAMING_SNAKE_CASE : int = 80_08_00 ): '''simple docstring''' lowerCAmelCase = degree * loga(SCREAMING_SNAKE_CASE ) lowerCAmelCase = int(SCREAMING_SNAKE_CASE ) lowerCAmelCase = calculate_prime_numbers(SCREAMING_SNAKE_CASE ) lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = len(SCREAMING_SNAKE_CASE ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(f'{solution() = }')
46
import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase (SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = DebertaTokenizer lowerCamelCase__ = True lowerCamelCase__ = DebertaTokenizerFast def __A ( self : List[Any] ) -> Dict: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE_ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "[UNK]", ] SCREAMING_SNAKE_CASE_ = dict(zip(__magic_name__ , range(len(__magic_name__ ) ) ) ) SCREAMING_SNAKE_CASE_ = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] SCREAMING_SNAKE_CASE_ = {"unk_token": "[UNK]"} SCREAMING_SNAKE_CASE_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) SCREAMING_SNAKE_CASE_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(__magic_name__ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(__magic_name__ ) ) def __A ( self : str , **__magic_name__ : int ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__magic_name__ ) def __A ( self : str , __magic_name__ : List[Any] ) -> Dict: SCREAMING_SNAKE_CASE_ = "lower newer" SCREAMING_SNAKE_CASE_ = "lower newer" return input_text, output_text def __A ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = "lower newer" SCREAMING_SNAKE_CASE_ = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] SCREAMING_SNAKE_CASE_ = tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = tokens + [tokenizer.unk_token] SCREAMING_SNAKE_CASE_ = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__magic_name__ ) , __magic_name__ ) def __A ( self : Optional[int] ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = tokenizer("Hello" , "World" ) SCREAMING_SNAKE_CASE_ = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd["token_type_ids"] , __magic_name__ ) @slow def __A ( self : Any ) -> Any: SCREAMING_SNAKE_CASE_ = self.tokenizer_class.from_pretrained("microsoft/deberta-base" ) SCREAMING_SNAKE_CASE_ = tokenizer.encode("sequence builders" , add_special_tokens=__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.encode("multi-sequence build" , add_special_tokens=__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.encode( "sequence builders" , add_special_tokens=__magic_name__ , add_prefix_space=__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=__magic_name__ , add_prefix_space=__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.build_inputs_with_special_tokens(__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.build_inputs_with_special_tokens(__magic_name__ , __magic_name__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def __A ( self : Tuple ) -> str: SCREAMING_SNAKE_CASE_ = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained("microsoft/deberta-base" ) SCREAMING_SNAKE_CASE_ = [ "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations", "ALBERT incorporates two parameter reduction techniques", "The first one is a factorized embedding parameterization. By decomposing the large vocabulary" " embedding matrix into two small matrices, we separate the size of the hidden layers from the size of" " vocabulary embedding.", ] SCREAMING_SNAKE_CASE_ = tokenizer(__magic_name__ , padding=__magic_name__ ) SCREAMING_SNAKE_CASE_ = [tokenizer.decode(__magic_name__ , skip_special_tokens=__magic_name__ ) for seq in encoding["input_ids"]] # fmt: off SCREAMING_SNAKE_CASE_ = { "input_ids": [ [1, 2_118, 11_126, 565, 35, 83, 25_191, 163, 18_854, 13, 12_156, 12, 16_101, 25_376, 13_807, 9, 22_205, 27_893, 1_635, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 2_118, 11_126, 565, 24_536, 80, 43_797, 4_878, 7_373, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 133, 78, 65, 16, 10, 3_724, 1_538, 33_183, 11_303, 43_797, 1_938, 4, 870, 24_165, 29_105, 5, 739, 32_644, 33_183, 11_303, 36_173, 88, 80, 650, 7_821, 45_940, 6, 52, 2_559, 5, 1_836, 9, 5, 7_397, 13_171, 31, 5, 1_836, 9, 32_644, 33_183, 11_303, 4, 2] ], "token_type_ids": [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], "attention_mask": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on SCREAMING_SNAKE_CASE_ = [ "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations", "ALBERT incorporates two parameter reduction techniques", "The first one is a factorized embedding parameterization. By decomposing the large vocabulary" " embedding matrix into two small matrices, we separate the size of the hidden layers from the size of" " vocabulary embedding.", ] self.assertDictEqual(encoding.data , __magic_name__ ) for expected, decoded in zip(__magic_name__ , __magic_name__ ): self.assertEqual(__magic_name__ , __magic_name__ )
118
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __A ( unittest.TestCase ): def __A ( self ): _lowerCAmelCase : str = """ZinengTang/tvlt-base""" _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() def __A ( self , **a__ ): return TvltImageProcessor.from_pretrained(self.checkpoint , **a__ ) def __A ( self , **a__ ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **a__ ) def __A ( self ): shutil.rmtree(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = self.get_image_processor() _lowerCAmelCase : Optional[int] = self.get_feature_extractor() _lowerCAmelCase : List[str] = TvltProcessor(image_processor=a__ , feature_extractor=a__ ) processor.save_pretrained(self.tmpdirname ) _lowerCAmelCase : str = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , a__ ) self.assertIsInstance(processor.image_processor , a__ ) def __A ( self ): _lowerCAmelCase : int = self.get_image_processor() _lowerCAmelCase : Union[str, Any] = self.get_feature_extractor() _lowerCAmelCase : List[str] = TvltProcessor(image_processor=a__ , feature_extractor=a__ ) _lowerCAmelCase : List[Any] = np.ones([12000] ) _lowerCAmelCase : Tuple = feature_extractor(a__ , return_tensors="""np""" ) _lowerCAmelCase : Dict = processor(audio=a__ , return_tensors="""np""" ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __A ( self ): _lowerCAmelCase : int = self.get_image_processor() _lowerCAmelCase : Tuple = self.get_feature_extractor() _lowerCAmelCase : List[Any] = TvltProcessor(image_processor=a__ , feature_extractor=a__ ) _lowerCAmelCase : Optional[Any] = np.ones([3, 224, 224] ) _lowerCAmelCase : Union[str, Any] = image_processor(a__ , return_tensors="""np""" ) _lowerCAmelCase : Tuple = processor(images=a__ , return_tensors="""np""" ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = self.get_image_processor() _lowerCAmelCase : Optional[int] = self.get_feature_extractor() _lowerCAmelCase : List[Any] = TvltProcessor(image_processor=a__ , feature_extractor=a__ ) _lowerCAmelCase : List[Any] = np.ones([12000] ) _lowerCAmelCase : Union[str, Any] = np.ones([3, 224, 224] ) _lowerCAmelCase : Tuple = processor(audio=a__ , images=a__ ) self.assertListEqual(list(inputs.keys() ) , ["""audio_values""", """audio_mask""", """pixel_values""", """pixel_mask"""] ) # test if it raises when no input is passed with pytest.raises(a__ ): processor() def __A ( self ): _lowerCAmelCase : List[Any] = self.get_image_processor() _lowerCAmelCase : List[str] = self.get_feature_extractor() _lowerCAmelCase : Tuple = TvltProcessor(image_processor=a__ , feature_extractor=a__ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg="""`processor` and `image_processor`+`feature_extractor` model input names do not match""" , )
352
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _a : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _a : Union[str, Any] = 250_004 _a : Optional[int] = 250_020 @require_sentencepiece @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Optional[Any] = MBartaaTokenizer _UpperCamelCase : Any = MBartaaTokenizerFast _UpperCamelCase : List[str] = True _UpperCamelCase : Optional[int] = True def __A ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : Tuple = MBartaaTokenizer(a__ , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=a__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : Any = """<s>""" _lowerCAmelCase : Optional[int] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a__ ) , a__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a__ ) , a__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(a__ ) , 1054 ) def __A ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1054 ) def __A ( self ): _lowerCAmelCase : str = MBartaaTokenizer(a__ , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=a__ ) _lowerCAmelCase : Optional[int] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : int = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a__ , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """."""] , ) _lowerCAmelCase : Union[str, Any] = tokenizer.convert_tokens_to_ids(a__ ) self.assertListEqual( a__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : int = tokenizer.convert_ids_to_tokens(a__ ) self.assertListEqual( a__ , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """."""] , ) @slow def __A ( self ): # fmt: off _lowerCAmelCase : Any = {"""input_ids""": [[250004, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [250004, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [250004, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a__ , model_name="""facebook/mbart-large-50""" , revision="""d3913889c59cd5c9e456b269c376325eabad57e2""" , ) def __A ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCAmelCase : List[str] = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-mbart50""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): _lowerCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(a__ , **a__ ) _lowerCAmelCase : List[str] = self.tokenizer_class.from_pretrained(a__ , **a__ ) _lowerCAmelCase : Optional[int] = tempfile.mkdtemp() _lowerCAmelCase : Optional[Any] = tokenizer_r.save_pretrained(a__ ) _lowerCAmelCase : List[str] = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) _lowerCAmelCase : Any = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : List[Any] = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : int = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=True _lowerCAmelCase : List[Any] = tempfile.mkdtemp() _lowerCAmelCase : Dict = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) _lowerCAmelCase : str = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : List[str] = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : Any = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=False _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Tuple = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) _lowerCAmelCase : str = tokenizer_p.save_pretrained(a__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCAmelCase : Any = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : int = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) @require_torch @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): _UpperCamelCase : Union[str, Any] = "facebook/mbart-large-50-one-to-many-mmt" _UpperCamelCase : int = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] _UpperCamelCase : Any = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] _UpperCamelCase : List[str] = [EN_CODE, 8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2] @classmethod def __A ( cls ): _lowerCAmelCase : MBartaaTokenizer = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en_XX""" , tgt_lang="""ro_RO""" ) _lowerCAmelCase : str = 1 return cls def __A ( self ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ar_AR"""] , 250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""en_EN"""] , 250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ro_RO"""] , 250020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""mr_IN"""] , 250038 ) def __A ( self ): _lowerCAmelCase : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , a__ ) def __A ( self ): self.assertIn(a__ , self.tokenizer.all_special_ids ) _lowerCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] _lowerCAmelCase : int = self.tokenizer.decode(a__ , skip_special_tokens=a__ ) _lowerCAmelCase : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=a__ ) self.assertEqual(a__ , a__ ) self.assertNotIn(self.tokenizer.eos_token , a__ ) def __A ( self ): _lowerCAmelCase : Any = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , a__ ) _lowerCAmelCase : Optional[Any] = 10 _lowerCAmelCase : Optional[int] = self.tokenizer(a__ , max_length=a__ , truncation=a__ ).input_ids[0] self.assertEqual(ids[0] , a__ ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(a__ ) , a__ ) def __A ( self ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [250053, 250001] ) def __A ( self ): _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Tuple = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(a__ ) _lowerCAmelCase : List[Any] = MBartaaTokenizer.from_pretrained(a__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , a__ ) @require_torch def __A ( self ): _lowerCAmelCase : List[str] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=a__ , return_tensors="""pt""" ) _lowerCAmelCase : Any = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def __A ( self ): _lowerCAmelCase : Optional[int] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) _lowerCAmelCase : Tuple = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) self.assertIsInstance(a__ , a__ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) _lowerCAmelCase : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , a__ ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def __A ( self ): _lowerCAmelCase : str = self.tokenizer(self.src_text , padding=a__ , truncation=a__ , max_length=3 , return_tensors="""pt""" ) _lowerCAmelCase : List[Any] = self.tokenizer( text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=10 , return_tensors="""pt""" ) _lowerCAmelCase : List[str] = targets["""input_ids"""] _lowerCAmelCase : Any = shift_tokens_right(a__ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def __A ( self ): _lowerCAmelCase : str = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""en_XX""" , tgt_lang="""ar_AR""" ) self.assertEqual( nested_simplify(a__ ) , { # en_XX, A, test, EOS """input_ids""": [[250004, 62, 3034, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 250001, } , )
126
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """microsoft/trocr-base-handwritten""": ( """https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json""" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = """trocr""" __lowerCAmelCase = ["""past_key_values"""] __lowerCAmelCase = { """num_attention_heads""": """decoder_attention_heads""", """hidden_size""": """d_model""", """num_hidden_layers""": """decoder_layers""", } def __init__( self : Optional[Any] , lowerCamelCase_ : Optional[int]=5_0265 , lowerCamelCase_ : Optional[int]=1024 , lowerCamelCase_ : List[Any]=12 , lowerCamelCase_ : Any=16 , lowerCamelCase_ : Tuple=4096 , lowerCamelCase_ : Tuple="gelu" , lowerCamelCase_ : List[str]=512 , lowerCamelCase_ : Union[str, Any]=0.1 , lowerCamelCase_ : List[str]=0.0 , lowerCamelCase_ : Optional[int]=0.0 , lowerCamelCase_ : Union[str, Any]=2 , lowerCamelCase_ : Tuple=0.0_2 , lowerCamelCase_ : Union[str, Any]=0.0 , lowerCamelCase_ : str=True , lowerCamelCase_ : List[Any]=False , lowerCamelCase_ : List[str]=True , lowerCamelCase_ : List[Any]=True , lowerCamelCase_ : List[str]=1 , lowerCamelCase_ : Optional[Any]=0 , lowerCamelCase_ : List[Any]=2 , **lowerCamelCase_ : Union[str, Any] , ): """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = d_model UpperCamelCase = decoder_layers UpperCamelCase = decoder_attention_heads UpperCamelCase = decoder_ffn_dim UpperCamelCase = activation_function UpperCamelCase = max_position_embeddings UpperCamelCase = dropout UpperCamelCase = attention_dropout UpperCamelCase = activation_dropout UpperCamelCase = init_std UpperCamelCase = decoder_layerdrop UpperCamelCase = use_cache UpperCamelCase = scale_embedding UpperCamelCase = use_learned_position_embeddings UpperCamelCase = layernorm_embedding super().__init__( pad_token_id=lowerCamelCase_ , bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , decoder_start_token_id=lowerCamelCase_ , **lowerCamelCase_ , )
343
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""" ) ) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue_model_parallelism.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_600, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_600, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, ] ) class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def lowerCamelCase_ ( self : Tuple ): """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="""utf-8""" , check=lowerCamelCase_ , ) assert hasattr(self , """env""" ) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : List[str] ): """simple docstring""" UpperCamelCase = { """enabled""": True, """processes_per_host""": 8, } UpperCamelCase = { """enabled""": True, """parameters""": { """microbatches""": 4, """placement_strategy""": """spread""", """pipeline""": """interleaved""", """optimize""": """speed""", """partitions""": 4, """ddp""": True, }, } UpperCamelCase = {"""smdistributed""": {"""modelparallel""": smp_options}, """mpi""": mpi_options} UpperCamelCase = """trainer""" if self.script == """run_glue.py""" else """smtrainer""" # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-{instance_count}-smp-{name_extension}""" , instance_count=lowerCamelCase_ , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase_ , hyperparameters={ **self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path, """max_steps""": 500, } , metric_definitions=self.env.metric_definitions , distribution=lowerCamelCase_ , py_version="""py36""" , ) def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : List[Any] ): """simple docstring""" TrainingJobAnalytics(lowerCamelCase_ ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(1,)] ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : int ): """simple docstring""" UpperCamelCase = self.create_estimator(lowerCamelCase_ ) # run training estimator.fit() # result dataframe UpperCamelCase = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCamelCase = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) UpperCamelCase = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCamelCase = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , lowerCamelCase_ )
343
1
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
351
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCamelCase = 16 _lowerCamelCase = 32 def a__ ( _SCREAMING_SNAKE_CASE : Accelerator , _SCREAMING_SNAKE_CASE : int = 16 ) -> List[str]: """simple docstring""" UpperCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained("bert-base-cased" ) UpperCAmelCase_ : Optional[Any] = load_dataset("glue" , "mrpc" ) def tokenize_function(_SCREAMING_SNAKE_CASE : int ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase_ : str = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase_ : Optional[Any] = datasets.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase_ : Any = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(_SCREAMING_SNAKE_CASE : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase_ : Optional[Any] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase_ : Optional[Any] = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase_ : Union[str, Any] = 8 else: UpperCAmelCase_ : List[str] = None return tokenizer.pad( _SCREAMING_SNAKE_CASE , padding="longest" , max_length=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_tensors="pt" , ) # Instantiate dataloaders. UpperCAmelCase_ : Union[str, Any] = DataLoader( tokenized_datasets["train"] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = DataLoader( tokenized_datasets["validation"] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowerCamelCase = mocked_dataloaders # noqa: F811 def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Tuple ) -> str: """simple docstring""" if os.environ.get("TESTING_MOCKED_DATALOADERS" , _SCREAMING_SNAKE_CASE ) == "1": UpperCAmelCase_ : Tuple = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: UpperCAmelCase_ : Optional[Any] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="all" , project_dir=args.project_dir ) else: UpperCAmelCase_ : List[Any] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase_ : Optional[Any] = config["lr"] UpperCAmelCase_ : Union[str, Any] = int(config["num_epochs"] ) UpperCAmelCase_ : str = int(config["seed"] ) UpperCAmelCase_ : Tuple = int(config["batch_size"] ) set_seed(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = get_dataloaders(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = evaluate.load("glue" , "mrpc" ) # If the batch size is too big we use gradient accumulation UpperCAmelCase_ : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: UpperCAmelCase_ : Tuple = batch_size // MAX_GPU_BATCH_SIZE UpperCAmelCase_ : Tuple = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase_ : Tuple = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=_SCREAMING_SNAKE_CASE ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase_ : Union[str, Any] = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase_ : int = AdamW(params=model.parameters() , lr=_SCREAMING_SNAKE_CASE ) # Instantiate scheduler UpperCAmelCase_ : Optional[int] = get_linear_schedule_with_warmup( optimizer=_SCREAMING_SNAKE_CASE , num_warmup_steps=1_00 , num_training_steps=(len(_SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : int = accelerator.prepare( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: UpperCAmelCase_ : List[str] = os.path.split(_SCREAMING_SNAKE_CASE )[-1].split("." )[0] accelerator.init_trackers(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Now we train the model for epoch in range(_SCREAMING_SNAKE_CASE ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: UpperCAmelCase_ : Dict = 0 for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) UpperCAmelCase_ : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() UpperCAmelCase_ : List[str] = loss / gradient_accumulation_steps accelerator.backward(_SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase_ : Optional[int] = model(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = outputs.logits.argmax(dim=-1 ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ : Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , _SCREAMING_SNAKE_CASE ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { "accuracy": eval_metric["accuracy"], "f1": eval_metric["f1"], "train_loss": total_loss.item() / len(_SCREAMING_SNAKE_CASE ), "epoch": epoch, } , step=_SCREAMING_SNAKE_CASE , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def a__ ( ) -> List[str]: """simple docstring""" UpperCAmelCase_ : int = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) parser.add_argument( "--with_tracking" , action="store_true" , help="Whether to load in all available experiment trackers from the environment and use them for logging." , ) parser.add_argument( "--project_dir" , type=_SCREAMING_SNAKE_CASE , default="logs" , help="Location on where to store experiment tracking logs` and relevent project information" , ) UpperCAmelCase_ : List[Any] = parser.parse_args() UpperCAmelCase_ : Dict = {"lr": 2E-5, "num_epochs": 3, "seed": 42, "batch_size": 16} training_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
67
0
'''simple docstring''' from __future__ import annotations from decimal import Decimal from numpy import array def UpperCamelCase ( _lowerCamelCase : list[list[float]] ): A__ = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(_lowerCamelCase ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix A__ = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creates a copy of the matrix with swapped positions of the elements A__ = [[0.0, 0.0], [0.0, 0.0]] A__, A__ = matrix[1][1], matrix[0][0] A__, A__ = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(_lowerCamelCase ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(_lowerCamelCase ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule A__ = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creating cofactor matrix A__ = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] A__ = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) A__ = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) A__ = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) A__ = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) A__ = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) A__ = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) A__ = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) A__ = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) A__ = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) A__ = array(_lowerCamelCase ) for i in range(3 ): for j in range(3 ): A__ = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix A__ = array(_lowerCamelCase ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(_lowerCamelCase ) # Calculate the inverse of the matrix return [[float(d(_lowerCamelCase ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError("Please provide a matrix of size 2x2 or 3x3." )
237
'''simple docstring''' import functools def UpperCamelCase ( _lowerCamelCase : str , _lowerCamelCase : str ): A__ = len(_lowerCamelCase ) A__ = len(_lowerCamelCase ) @functools.cache def min_distance(_lowerCamelCase : int , _lowerCamelCase : int ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa A__ = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , _lowerCamelCase ) , 1 + min_distance(_lowerCamelCase , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
237
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available A_ : List[str] ={"""configuration_speech_encoder_decoder""": ["""SpeechEncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] =["""SpeechEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] =["""FlaxSpeechEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys A_ : str =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
"""simple docstring""" from math import factorial, pi def SCREAMING_SNAKE_CASE_ ( snake_case : float , snake_case : int = 30 )-> float: if not isinstance(snake_case , (int, float) ): raise ValueError('maclaurin_sin() requires either an int or float for theta' ) if not isinstance(snake_case , snake_case ) or accuracy <= 0: raise ValueError('maclaurin_sin() requires a positive int for accuracy' ) _lowerCamelCase = float(snake_case ) _lowerCamelCase = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(snake_case ) ) def SCREAMING_SNAKE_CASE_ ( snake_case : float , snake_case : int = 30 )-> float: if not isinstance(snake_case , (int, float) ): raise ValueError('maclaurin_cos() requires either an int or float for theta' ) if not isinstance(snake_case , snake_case ) or accuracy <= 0: raise ValueError('maclaurin_cos() requires a positive int for accuracy' ) _lowerCamelCase = float(snake_case ) _lowerCamelCase = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(snake_case ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(1_0)) print(maclaurin_sin(-1_0)) print(maclaurin_sin(1_0, 1_5)) print(maclaurin_sin(-1_0, 1_5)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(1_0, 1_5)) print(maclaurin_cos(-1_0, 1_5))
80
1
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case_ ( self : Optional[int] , _snake_case : Any ): __lowercase : Optional[Any] = 3 __lowercase : Union[str, Any] = 250 __lowercase : Union[str, Any] = ids_tensor((batch_size, length) , _snake_case ) __lowercase : List[Any] = torch.ones((batch_size, length) , device=_snake_case , dtype=torch.float ) / length return input_ids, scores def snake_case_ ( self : Optional[Any] ): __lowercase : Optional[Any] = self._get_tensors(5 ) __lowercase : int = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Optional[Any] = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Any = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case_ ( self : str ): __lowercase : Optional[int] = MaxLengthCriteria(max_length=10 ) __lowercase : Dict = self._get_tensors(5 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Tuple = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : str = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case_ ( self : List[Any] ): __lowercase : Any = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) __lowercase : Optional[int] = self._get_tensors(5 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Dict = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Tuple = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) __lowercase : Optional[int] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def snake_case_ ( self : Optional[Any] ): __lowercase : List[Any] = self._get_tensors(5 ) __lowercase : Dict = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) __lowercase : Dict = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case_ ( self : Dict ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(_snake_case ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) __lowercase : Tuple = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(_snake_case ) , 1 )
156
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
0
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig lowercase : str = logging.get_logger(__name__) class A__ : """simple docstring""" def __init__( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__ : Optional[Any] = question_encoder a__ : Optional[int] = generator a__ : str = self.question_encoder def __lowercase ( self , lowercase) -> Dict: '''simple docstring''' if os.path.isfile(lowercase): raise ValueError(F'Provided path ({save_directory}) should be a directory, not a file') os.makedirs(lowercase , exist_ok=lowercase) a__ : List[Any] = os.path.join(lowercase , 'question_encoder_tokenizer') a__ : List[Any] = os.path.join(lowercase , 'generator_tokenizer') self.question_encoder.save_pretrained(lowercase) self.generator.save_pretrained(lowercase) @classmethod def __lowercase ( cls , lowercase , **lowercase) -> Any: '''simple docstring''' from ..auto.tokenization_auto import AutoTokenizer a__ : Optional[int] = kwargs.pop('config' , lowercase) if config is None: a__ : Optional[int] = RagConfig.from_pretrained(lowercase) a__ : Union[str, Any] = AutoTokenizer.from_pretrained( lowercase , config=config.question_encoder , subfolder='question_encoder_tokenizer') a__ : Optional[int] = AutoTokenizer.from_pretrained( lowercase , config=config.generator , subfolder='generator_tokenizer') return cls(question_encoder=lowercase , generator=lowercase) def __call__( self , *lowercase , **lowercase) -> Any: '''simple docstring''' return self.current_tokenizer(*lowercase , **lowercase) def __lowercase ( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' return self.generator.batch_decode(*lowercase , **lowercase) def __lowercase ( self , *lowercase , **lowercase) -> Optional[Any]: '''simple docstring''' return self.generator.decode(*lowercase , **lowercase) def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : Union[str, Any] = self.question_encoder def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : Optional[int] = self.generator def __lowercase ( self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = "longest" , lowercase = None , lowercase = True , **lowercase , ) -> BatchEncoding: '''simple docstring''' warnings.warn( '`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ' 'regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ' 'context manager to prepare your targets. See the documentation of your specific tokenizer for more ' 'details' , lowercase , ) if max_length is None: a__ : Union[str, Any] = self.current_tokenizer.model_max_length a__ : Optional[Any] = self( lowercase , add_special_tokens=lowercase , return_tensors=lowercase , max_length=lowercase , padding=lowercase , truncation=lowercase , **lowercase , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: a__ : Any = self.current_tokenizer.model_max_length a__ : Tuple = self( text_target=lowercase , add_special_tokens=lowercase , return_tensors=lowercase , padding=lowercase , max_length=lowercase , truncation=lowercase , **lowercase , ) a__ : List[str] = labels['input_ids'] return model_inputs
225
import pickle import numpy as np from matplotlib import pyplot as plt class A__ : """simple docstring""" def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=0.2 , lowercase=0.2) -> Any: '''simple docstring''' a__ : Tuple = bp_numa a__ : Union[str, Any] = bp_numa a__ : Optional[int] = bp_numa a__ : Optional[int] = conva_get[:2] a__ : Optional[Any] = conva_get[2] a__ : Optional[int] = size_pa a__ : Union[str, Any] = rate_w a__ : Dict = rate_t a__ : int = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0]) + 0.5) for i in range(self.conva[1]) ] a__ : Optional[int] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa) + 0.5) a__ : Optional[int] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa) + 0.5) a__ : Any = -2 * np.random.rand(self.conva[1]) + 1 a__ : Optional[Any] = -2 * np.random.rand(self.num_bpa) + 1 a__ : Optional[int] = -2 * np.random.rand(self.num_bpa) + 1 def __lowercase ( self , lowercase) -> List[Any]: '''simple docstring''' a__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(lowercase , 'wb') as f: pickle.dump(lowercase , lowercase) print(F'Model saved: {save_path}') @classmethod def __lowercase ( cls , lowercase) -> Any: '''simple docstring''' with open(lowercase , 'rb') as f: a__ : Any = pickle.load(lowercase) # noqa: S301 a__ : Dict = model_dic.get('conv1') conv_get.append(model_dic.get('step_conv1')) a__ : Tuple = model_dic.get('size_pooling1') a__ : Optional[int] = model_dic.get('num_bp1') a__ : Tuple = model_dic.get('num_bp2') a__ : Optional[Any] = model_dic.get('num_bp3') a__ : Optional[Any] = model_dic.get('rate_weight') a__ : int = model_dic.get('rate_thre') # create model instance a__ : Union[str, Any] = CNN(lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) # modify model parameter a__ : str = model_dic.get('w_conv1') a__ : Optional[int] = model_dic.get('wkj') a__ : Tuple = model_dic.get('vji') a__ : str = model_dic.get('thre_conv1') a__ : List[str] = model_dic.get('thre_bp2') a__ : Tuple = model_dic.get('thre_bp3') return conv_ins def __lowercase ( self , lowercase) -> Any: '''simple docstring''' return 1 / (1 + np.exp(-1 * x)) def __lowercase ( self , lowercase) -> Optional[int]: '''simple docstring''' return round(lowercase , 3) def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase) -> Any: '''simple docstring''' a__ : Union[str, Any] = convs[0] a__ : Tuple = convs[1] a__ : Any = np.shape(lowercase)[0] # get the data slice of original image data, data_focus a__ : Tuple = [] for i_focus in range(0 , size_data - size_conv + 1 , lowercase): for j_focus in range(0 , size_data - size_conv + 1 , lowercase): a__ : str = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(lowercase) # calculate the feature map of every single kernel, and saved as list of matrix a__ : str = [] a__ : Union[str, Any] = int((size_data - size_conv) / conv_step + 1) for i_map in range(lowercase): a__ : Tuple = [] for i_focus in range(len(lowercase)): a__ : Optional[Any] = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map])) - thre_convs[i_map] ) featuremap.append(self.sig(lowercase)) a__ : Dict = np.asmatrix(lowercase).reshape( lowercase , lowercase) data_featuremap.append(lowercase) # expanding the data slice to One dimenssion a__ : int = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(lowercase)) a__ : Optional[int] = np.asarray(lowercase) return focus_list, data_featuremap def __lowercase ( self , lowercase , lowercase , lowercase="average_pool") -> str: '''simple docstring''' a__ : Any = len(featuremaps[0]) a__ : int = int(size_map / size_pooling) a__ : Optional[Any] = [] for i_map in range(len(lowercase)): a__ : Any = featuremaps[i_map] a__ : Optional[int] = [] for i_focus in range(0 , lowercase , lowercase): for j_focus in range(0 , lowercase , lowercase): a__ : Any = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(lowercase)) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(lowercase)) a__ : List[str] = np.asmatrix(lowercase).reshape(lowercase , lowercase) featuremap_pooled.append(lowercase) return featuremap_pooled def __lowercase ( self , lowercase) -> Optional[Any]: '''simple docstring''' a__ : Any = [] for i in range(len(lowercase)): a__ : Tuple = np.shape(data[i]) a__ : List[str] = data[i].reshape(1 , shapes[0] * shapes[1]) a__ : Optional[Any] = data_listed.getA().tolist()[0] data_expanded.extend(lowercase) a__ : Union[str, Any] = np.asarray(lowercase) return data_expanded def __lowercase ( self , lowercase) -> Dict: '''simple docstring''' a__ : Dict = np.asarray(lowercase) a__ : Optional[int] = np.shape(lowercase) a__ : Any = data_mat.reshape(1 , shapes[0] * shapes[1]) return data_expanded def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' a__ : int = [] a__ : Optional[int] = 0 for i_map in range(lowercase): a__ : Optional[Any] = np.ones((size_map, size_map)) for i in range(0 , lowercase , lowercase): for j in range(0 , lowercase , lowercase): a__ : Union[str, Any] = pd_pool[ i_pool ] a__ : Tuple = i_pool + 1 a__ : Optional[int] = np.multiply( lowercase , np.multiply(out_map[i_map] , (1 - out_map[i_map]))) pd_all.append(lowercase) return pd_all def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=bool) -> str: '''simple docstring''' print('----------------------Start Training-------------------------') print((' - - Shape: Train_Data ', np.shape(lowercase))) print((' - - Shape: Teach_Data ', np.shape(lowercase))) a__ : Dict = 0 a__ : List[Any] = [] a__ : Optional[int] = 1_0000 while rp < n_repeat and mse >= error_accuracy: a__ : Dict = 0 print(F'-------------Learning Time {rp}--------------') for p in range(len(lowercase)): # print('------------Learning Image: %d--------------'%p) a__ : Dict = np.asmatrix(datas_train[p]) a__ : Any = np.asarray(datas_teach[p]) a__ , a__ : Optional[int] = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) a__ : Dict = self.pooling(lowercase , self.size_poolinga) a__ : Optional[Any] = np.shape(lowercase) a__ : Union[str, Any] = self._expand(lowercase) a__ : List[Any] = data_bp_input a__ : Tuple = np.dot(lowercase , self.vji.T) - self.thre_bpa a__ : Any = self.sig(lowercase) a__ : Any = np.dot(lowercase , self.wkj.T) - self.thre_bpa a__ : Any = self.sig(lowercase) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- a__ : Any = np.multiply( (data_teach - bp_outa) , np.multiply(lowercase , (1 - bp_outa))) a__ : Optional[Any] = np.multiply( np.dot(lowercase , self.wkj) , np.multiply(lowercase , (1 - bp_outa))) a__ : Tuple = np.dot(lowercase , self.vji) a__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) a__ : List[str] = pd_conva_pooled.T.getA().tolist() a__ : str = self._calculate_gradient_from_pool( lowercase , lowercase , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1]): a__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv]) a__ : int = self.rate_weight * np.dot(lowercase , lowercase) a__ : List[str] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0])) a__ : str = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv]) * self.rate_thre ) # all connected layer a__ : List[str] = self.wkj + pd_k_all.T * bp_outa * self.rate_weight a__ : List[str] = self.vji + pd_j_all.T * bp_outa * self.rate_weight a__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre a__ : Tuple = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image a__ : List[str] = np.sum(abs(data_teach - bp_outa)) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) a__ : Any = rp + 1 a__ : Optional[Any] = error_count / patterns all_mse.append(lowercase) def draw_error(): a__ : int = [error_accuracy for i in range(int(n_repeat * 1.2))] plt.plot(lowercase , '+-') plt.plot(lowercase , 'r--') plt.xlabel('Learning Times') plt.ylabel('All_mse') plt.grid(lowercase , alpha=0.5) plt.show() print('------------------Training Complished---------------------') print((' - - Training epoch: ', rp, F' - - Mse: {mse:.6f}')) if draw_e: draw_error() return mse def __lowercase ( self , lowercase) -> Optional[int]: '''simple docstring''' a__ : str = [] print('-------------------Start Testing-------------------------') print((' - - Shape: Test_Data ', np.shape(lowercase))) for p in range(len(lowercase)): a__ : int = np.asmatrix(datas_test[p]) a__ , a__ : Optional[int] = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) a__ : str = self.pooling(lowercase , self.size_poolinga) a__ : Optional[int] = self._expand(lowercase) a__ : str = data_bp_input a__ : Union[str, Any] = bp_outa * self.vji.T - self.thre_bpa a__ : Optional[Any] = self.sig(lowercase) a__ : int = bp_outa * self.wkj.T - self.thre_bpa a__ : Dict = self.sig(lowercase) produce_out.extend(bp_outa.getA().tolist()) a__ : List[Any] = [list(map(self.do_round , lowercase)) for each in produce_out] return np.asarray(lowercase) def __lowercase ( self , lowercase) -> str: '''simple docstring''' a__ : str = np.asmatrix(lowercase) a__ , a__ : str = self.convolute( lowercase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) a__ : List[str] = self.pooling(lowercase , self.size_poolinga) return data_conveda, data_pooleda if __name__ == "__main__": pass
225
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer A_ :Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ :Dict = { '''vocab_file''': { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''unc-nlp/lxmert-base-uncased''': ( '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ :str = { '''unc-nlp/lxmert-base-uncased''': 512, } A_ :Dict = { '''unc-nlp/lxmert-base-uncased''': {'''do_lower_case''': True}, } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[Any] =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Union[str, Any] =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : str =LxmertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : Dict =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCamelCase__ ) != tokenize_chinese_chars ): __UpperCamelCase : str =getattr(lowerCamelCase__ , normalizer_state.pop('type' ) ) __UpperCamelCase : Any =do_lower_case __UpperCamelCase : Dict =strip_accents __UpperCamelCase : List[str] =tokenize_chinese_chars __UpperCamelCase : Union[str, Any] =normalizer_class(**lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =do_lower_case def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" __UpperCamelCase : List[str] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : str =[self.sep_token_id] __UpperCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : int =self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
71
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class __A ( unittest.TestCase , UpperCamelCase__ ): def _lowercase (self : Tuple ): UpperCAmelCase_ = load_tool("text-to-speech" ) self.tool.setup() def _lowercase (self : Union[str, Any] ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) UpperCAmelCase_ = self.tool("hey" ) UpperCAmelCase_ = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) ) def _lowercase (self : List[str] ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) UpperCAmelCase_ = self.tool("hey" ) UpperCAmelCase_ = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) )
352
'''simple docstring''' import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch SCREAMING_SNAKE_CASE_: Dict =True except ImportError: SCREAMING_SNAKE_CASE_: str =False try: from torch.hub import _get_torch_home SCREAMING_SNAKE_CASE_: Optional[Any] =_get_torch_home() except ImportError: SCREAMING_SNAKE_CASE_: Union[str, Any] =os.path.expanduser( os.getenv('TORCH_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'torch')) ) SCREAMING_SNAKE_CASE_: int =os.path.join(torch_cache_home, 'transformers') SCREAMING_SNAKE_CASE_: Tuple ='https://cdn.huggingface.co' SCREAMING_SNAKE_CASE_: str ='https://s3.amazonaws.com/models.huggingface.co/bert' SCREAMING_SNAKE_CASE_: str ='/'.join(str(Path(__file__).resolve()).split('/')[:-1]) SCREAMING_SNAKE_CASE_: Optional[Any] =os.path.join(PATH, 'config.yaml') SCREAMING_SNAKE_CASE_: Optional[Any] =os.path.join(PATH, 'attributes.txt') SCREAMING_SNAKE_CASE_: Any =os.path.join(PATH, 'objects.txt') SCREAMING_SNAKE_CASE_: Optional[int] =os.getenv('PYTORCH_PRETRAINED_BERT_CACHE', default_cache_path) SCREAMING_SNAKE_CASE_: int =os.getenv('PYTORCH_TRANSFORMERS_CACHE', PYTORCH_PRETRAINED_BERT_CACHE) SCREAMING_SNAKE_CASE_: List[str] =os.getenv('TRANSFORMERS_CACHE', PYTORCH_TRANSFORMERS_CACHE) SCREAMING_SNAKE_CASE_: str ='pytorch_model.bin' SCREAMING_SNAKE_CASE_: Dict ='config.yaml' def lowerCAmelCase_ ( snake_case_ : Optional[int]=OBJECTS , snake_case_ : Optional[Any]=ATTRIBUTES ) -> Any: '''simple docstring''' UpperCAmelCase_ = [] with open(snake_case_ ) as f: for object in f.readlines(): vg_classes.append(object.split("," )[0].lower().strip() ) UpperCAmelCase_ = [] with open(snake_case_ ) as f: for object in f.readlines(): vg_attrs.append(object.split("," )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCAmelCase_ ( snake_case_ : Optional[int] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = OrderedDict() with open(snake_case_ , "rb" ) as f: UpperCAmelCase_ = pkl.load(snake_case_ )["model"] for k in copy.deepcopy(list(ckp.keys() ) ): UpperCAmelCase_ = ckp.pop(snake_case_ ) if isinstance(snake_case_ , np.ndarray ): UpperCAmelCase_ = torch.tensor(snake_case_ ) else: assert isinstance(snake_case_ , torch.tensor ), type(snake_case_ ) UpperCAmelCase_ = v return r class __A : a__ : Optional[Any] = {} def __init__(self : Union[str, Any] , __a : dict , __a : str = "root" , __a : str=0 ): UpperCAmelCase_ = name UpperCAmelCase_ = level UpperCAmelCase_ = {} for k, v in dictionary.items(): if v is None: raise ValueError() UpperCAmelCase_ = copy.deepcopy(__a ) UpperCAmelCase_ = copy.deepcopy(__a ) if isinstance(__a , __a ): UpperCAmelCase_ = Config(__a , name=__a , level=level + 1 ) UpperCAmelCase_ = v setattr(self , __a , __a ) UpperCAmelCase_ = d def __repr__(self : List[Any] ): return str(list((self._pointer.keys()) ) ) def __setattr__(self : int , __a : str , __a : Dict ): UpperCAmelCase_ = val UpperCAmelCase_ = val UpperCAmelCase_ = key.split("." ) UpperCAmelCase_ = len(__a ) - 1 UpperCAmelCase_ = self._pointer if len(__a ) > 1: for i, l in enumerate(__a ): if hasattr(self , __a ) and isinstance(getattr(self , __a ) , __a ): setattr(getattr(self , __a ) , ".".join(levels[i:] ) , __a ) if l == last_level: UpperCAmelCase_ = val else: UpperCAmelCase_ = pointer[l] def _lowercase (self : Optional[Any] ): return self._pointer def _lowercase (self : int , __a : Union[str, Any] , __a : str ): with open(f"""{file_name}""" , "w" ) as stream: dump(__a , __a ) def _lowercase (self : Any , __a : Optional[Any] , __a : List[str] ): with open(f"""{file_name}""" , "w" ) as stream: json.dump(__a , __a ) @staticmethod def _lowercase (__a : str ): with open(__a ) as stream: UpperCAmelCase_ = load(__a , Loader=__a ) return data def __str__(self : Dict ): UpperCAmelCase_ = " " if self._name != "root": UpperCAmelCase_ = f"""{t * (self._level-1)}{self._name}:\n""" else: UpperCAmelCase_ = "" UpperCAmelCase_ = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__a , __a ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__a ).__name__})\n""" UpperCAmelCase_ = level return r[:-1] @classmethod def _lowercase (cls : Tuple , __a : str , **__a : Dict ): UpperCAmelCase_ , UpperCAmelCase_ = cls.get_config_dict(__a , **__a ) return cls(__a ) @classmethod def _lowercase (cls : Any , __a : str , **__a : Dict ): UpperCAmelCase_ = kwargs.pop("cache_dir" , __a ) UpperCAmelCase_ = kwargs.pop("force_download" , __a ) UpperCAmelCase_ = kwargs.pop("resume_download" , __a ) UpperCAmelCase_ = kwargs.pop("proxies" , __a ) UpperCAmelCase_ = kwargs.pop("local_files_only" , __a ) if os.path.isdir(__a ): UpperCAmelCase_ = os.path.join(__a , __a ) elif os.path.isfile(__a ) or is_remote_url(__a ): UpperCAmelCase_ = pretrained_model_name_or_path else: UpperCAmelCase_ = hf_bucket_url(__a , filename=__a , use_cdn=__a ) try: # Load from URL or cache if already cached UpperCAmelCase_ = cached_path( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , ) # Load config dict if resolved_config_file is None: raise EnvironmentError UpperCAmelCase_ = Config.load_yaml(__a ) except EnvironmentError: UpperCAmelCase_ = "Can't load config for" raise EnvironmentError(__a ) if resolved_config_file == config_file: print("loading configuration file from path" ) else: print("loading configuration file cache" ) return Config.load_yaml(__a ), kwargs def lowerCAmelCase_ ( snake_case_ : str ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = torch.load("dump.pt" , map_location=in_tensor.device ) UpperCAmelCase_ = in_tensor.numpy() UpperCAmelCase_ = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(snake_case_ , snake_case_ , rtol=0.01 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(snake_case_ , snake_case_ , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*1_00:.4f} %""" " element-wise mismatch" ) raise Exception("tensors are all good" ) # Hugging face functions below def lowerCAmelCase_ ( snake_case_ : Optional[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = urlparse(snake_case_ ) return parsed.scheme in ("http", "https") def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : str , snake_case_ : Optional[int]=True ) -> str: '''simple docstring''' UpperCAmelCase_ = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX UpperCAmelCase_ = "/" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Union[str, Any] , snake_case_ : Optional[int]=None , snake_case_ : List[Any]=0 , snake_case_ : int=None , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = "python/{}".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(snake_case_ , snake_case_ ): ua += "; " + "; ".join("{}/{}".format(snake_case_ , snake_case_ ) for k, v in user_agent.items() ) elif isinstance(snake_case_ , snake_case_ ): ua += "; " + user_agent UpperCAmelCase_ = {"user-agent": ua} if resume_size > 0: UpperCAmelCase_ = "bytes=%d-" % (resume_size,) UpperCAmelCase_ = requests.get(snake_case_ , stream=snake_case_ , proxies=snake_case_ , headers=snake_case_ ) if response.status_code == 4_16: # Range not satisfiable return UpperCAmelCase_ = response.headers.get("Content-Length" ) UpperCAmelCase_ = resume_size + int(snake_case_ ) if content_length is not None else None UpperCAmelCase_ = tqdm( unit="B" , unit_scale=snake_case_ , total=snake_case_ , initial=snake_case_ , desc="Downloading" , ) for chunk in response.iter_content(chunk_size=10_24 ): if chunk: # filter out keep-alive new chunks progress.update(len(snake_case_ ) ) temp_file.write(snake_case_ ) progress.close() def lowerCAmelCase_ ( snake_case_ : Optional[int] , snake_case_ : str=None , snake_case_ : List[str]=False , snake_case_ : List[str]=None , snake_case_ : int=10 , snake_case_ : Any=False , snake_case_ : int=None , snake_case_ : str=False , ) -> str: '''simple docstring''' if cache_dir is None: UpperCAmelCase_ = TRANSFORMERS_CACHE if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) UpperCAmelCase_ = None if not local_files_only: try: UpperCAmelCase_ = requests.head(snake_case_ , allow_redirects=snake_case_ , proxies=snake_case_ , timeout=snake_case_ ) if response.status_code == 2_00: UpperCAmelCase_ = response.headers.get("ETag" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass UpperCAmelCase_ = url_to_filename(snake_case_ , snake_case_ ) # get cache path to put the file UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(snake_case_ ): return cache_path else: UpperCAmelCase_ = [ file for file in fnmatch.filter(os.listdir(snake_case_ ) , filename + ".*" ) if not file.endswith(".json" ) and not file.endswith(".lock" ) ] if len(snake_case_ ) > 0: return os.path.join(snake_case_ , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( "Cannot find the requested files in the cached path and outgoing traffic has been" " disabled. To enable model look-ups and downloads online, set 'local_files_only'" " to False." ) return None # From now on, etag is not None. if os.path.exists(snake_case_ ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. UpperCAmelCase_ = cache_path + ".lock" with FileLock(snake_case_ ): # If the download just completed while the lock was activated. if os.path.exists(snake_case_ ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: UpperCAmelCase_ = cache_path + ".incomplete" @contextmanager def _resumable_file_manager(): with open(snake_case_ , "a+b" ) as f: yield f UpperCAmelCase_ = _resumable_file_manager if os.path.exists(snake_case_ ): UpperCAmelCase_ = os.stat(snake_case_ ).st_size else: UpperCAmelCase_ = 0 else: UpperCAmelCase_ = partial(tempfile.NamedTemporaryFile , dir=snake_case_ , delete=snake_case_ ) UpperCAmelCase_ = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( "%s not found in cache or force_download set to True, downloading to %s" , snake_case_ , temp_file.name , ) http_get( snake_case_ , snake_case_ , proxies=snake_case_ , resume_size=snake_case_ , user_agent=snake_case_ , ) os.replace(temp_file.name , snake_case_ ) UpperCAmelCase_ = {"url": url, "etag": etag} UpperCAmelCase_ = cache_path + ".json" with open(snake_case_ , "w" ) as meta_file: json.dump(snake_case_ , snake_case_ ) return cache_path def lowerCAmelCase_ ( snake_case_ : Optional[Any] , snake_case_ : Any=None ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = url.encode("utf-8" ) UpperCAmelCase_ = shaaaa(snake_case_ ) UpperCAmelCase_ = url_hash.hexdigest() if etag: UpperCAmelCase_ = etag.encode("utf-8" ) UpperCAmelCase_ = shaaaa(snake_case_ ) filename += "." + etag_hash.hexdigest() if url.endswith(".h5" ): filename += ".h5" return filename def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : Tuple=None , snake_case_ : int=False , snake_case_ : Any=None , snake_case_ : List[Any]=False , snake_case_ : Any=None , snake_case_ : Any=False , snake_case_ : List[str]=False , snake_case_ : str=False , ) -> Union[str, Any]: '''simple docstring''' if cache_dir is None: UpperCAmelCase_ = TRANSFORMERS_CACHE if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) if is_remote_url(snake_case_ ): # URL, so get it from the cache (downloading if necessary) UpperCAmelCase_ = get_from_cache( snake_case_ , cache_dir=snake_case_ , force_download=snake_case_ , proxies=snake_case_ , resume_download=snake_case_ , user_agent=snake_case_ , local_files_only=snake_case_ , ) elif os.path.exists(snake_case_ ): # File, and it exists. UpperCAmelCase_ = url_or_filename elif urlparse(snake_case_ ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("file {} not found".format(snake_case_ ) ) else: # Something unknown raise ValueError("unable to parse {} as a URL or as a local path".format(snake_case_ ) ) if extract_compressed_file: if not is_zipfile(snake_case_ ) and not tarfile.is_tarfile(snake_case_ ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" UpperCAmelCase_ , UpperCAmelCase_ = os.path.split(snake_case_ ) UpperCAmelCase_ = output_file.replace("." , "-" ) + "-extracted" UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) if os.path.isdir(snake_case_ ) and os.listdir(snake_case_ ) and not force_extract: return output_path_extracted # Prevent parallel extractions UpperCAmelCase_ = output_path + ".lock" with FileLock(snake_case_ ): shutil.rmtree(snake_case_ , ignore_errors=snake_case_ ) os.makedirs(snake_case_ ) if is_zipfile(snake_case_ ): with ZipFile(snake_case_ , "r" ) as zip_file: zip_file.extractall(snake_case_ ) zip_file.close() elif tarfile.is_tarfile(snake_case_ ): UpperCAmelCase_ = tarfile.open(snake_case_ ) tar_file.extractall(snake_case_ ) tar_file.close() else: raise EnvironmentError("Archive format of {} could not be identified".format(snake_case_ ) ) return output_path_extracted return output_path def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : Optional[int]="," ) -> int: '''simple docstring''' assert isinstance(snake_case_ , snake_case_ ) if os.path.isfile(snake_case_ ): with open(snake_case_ ) as f: UpperCAmelCase_ = eval(f.read() ) else: UpperCAmelCase_ = requests.get(snake_case_ ) try: UpperCAmelCase_ = requests.json() except Exception: UpperCAmelCase_ = req.content.decode() assert data is not None, "could not connect" try: UpperCAmelCase_ = eval(snake_case_ ) except Exception: UpperCAmelCase_ = data.split("\n" ) req.close() return data def lowerCAmelCase_ ( snake_case_ : List[str] ) -> Any: '''simple docstring''' UpperCAmelCase_ = requests.get(snake_case_ ) UpperCAmelCase_ = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCAmelCase_ ( snake_case_ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = url.split("/" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(snake_case_ ) with open(snake_case_ , "rb" ) as stream: UpperCAmelCase_ = pkl.load(snake_case_ ) UpperCAmelCase_ = weights.pop("model" ) UpperCAmelCase_ = {} for k, v in model.items(): UpperCAmelCase_ = torch.from_numpy(snake_case_ ) if "running_var" in k: UpperCAmelCase_ = torch.tensor([0] ) UpperCAmelCase_ = k.replace("running_var" , "num_batches_tracked" ) UpperCAmelCase_ = zero return new def lowerCAmelCase_ ( ) -> int: '''simple docstring''' print(f"""{os.path.abspath(os.path.join(snake_case_ , os.pardir ) )}/demo.ipynb""" ) def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : Any="RGB" ) -> Dict: '''simple docstring''' assert isinstance(snake_case_ , snake_case_ ) if os.path.isfile(snake_case_ ): UpperCAmelCase_ = cva.imread(snake_case_ ) else: UpperCAmelCase_ = get_image_from_url(snake_case_ ) assert img is not None, f"""could not connect to: {im}""" UpperCAmelCase_ = cva.cvtColor(snake_case_ , cva.COLOR_BGR2RGB ) if input_format == "RGB": UpperCAmelCase_ = img[:, :, ::-1] return img def lowerCAmelCase_ ( snake_case_ : Tuple , snake_case_ : Union[str, Any]=1 ) -> str: '''simple docstring''' return (images[i : i + batch] for i in range(0 , len(snake_case_ ) , snake_case_ ))
106
0
import math def lowerCamelCase__ ( snake_case_ : int ) -> bool: assert isinstance(snake_case_ , snake_case_ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False __snake_case = range(3 , int(math.sqrt(snake_case_ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def lowerCamelCase__ ( snake_case_ : Optional[Any] , snake_case_ : Dict=1 , **snake_case_ : List[Any] ) -> str: __snake_case = factor * value __snake_case = value while not is_prime(snake_case_ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **snake_case_ ) return value
24
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class UpperCAmelCase_ ( __lowercase ): def __lt__( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> List[Any]: return self[-1] < other[-1] def __eq__( self : str , UpperCAmelCase__ : List[str] ) -> Tuple: return self[-1] == other[-1] def a_ ( lowerCamelCase : list ): lowerCAmelCase = [] # sort into stacks for element in collection: lowerCAmelCase = Stack([element] ) lowerCAmelCase = bisect_left(lowerCamelCase , lowerCamelCase ) if i != len(lowerCamelCase ): stacks[i].append(lowerCamelCase ) else: stacks.append(lowerCamelCase ) # use a heap-based merge to merge stack efficiently lowerCAmelCase = merge(*(reversed(lowerCamelCase ) for stack in stacks) ) return collection if __name__ == "__main__": __snake_case =input("""Enter numbers separated by a comma:\n""").strip() __snake_case =[int(item) for item in user_input.split(""",""")] print(patience_sort(unsorted))
4
0
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Any = { "facebook/detr-resnet-50": "https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json", # See all DETR models at https://huggingface.co/models?filter=detr } class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = 'detr' __magic_name__ = ['past_key_values'] __magic_name__ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , __snake_case=True , __snake_case=None , __snake_case=3 , __snake_case=1_0_0 , __snake_case=6 , __snake_case=2_0_4_8 , __snake_case=8 , __snake_case=6 , __snake_case=2_0_4_8 , __snake_case=8 , __snake_case=0.0 , __snake_case=0.0 , __snake_case=True , __snake_case="relu" , __snake_case=2_5_6 , __snake_case=0.1 , __snake_case=0.0 , __snake_case=0.0 , __snake_case=0.02 , __snake_case=1.0 , __snake_case=False , __snake_case="sine" , __snake_case="resnet50" , __snake_case=True , __snake_case=False , __snake_case=1 , __snake_case=5 , __snake_case=2 , __snake_case=1 , __snake_case=1 , __snake_case=5 , __snake_case=2 , __snake_case=0.1 , **__snake_case , ): if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) snake_case = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(__snake_case , __snake_case ): snake_case = backbone_config.get('''model_type''' ) snake_case = CONFIG_MAPPING[backbone_model_type] snake_case = config_class.from_dict(__snake_case ) # set timm attributes to None snake_case , snake_case , snake_case = None, None, None snake_case = use_timm_backbone snake_case = backbone_config snake_case = num_channels snake_case = num_queries snake_case = d_model snake_case = encoder_ffn_dim snake_case = encoder_layers snake_case = encoder_attention_heads snake_case = decoder_ffn_dim snake_case = decoder_layers snake_case = decoder_attention_heads snake_case = dropout snake_case = attention_dropout snake_case = activation_dropout snake_case = activation_function snake_case = init_std snake_case = init_xavier_std snake_case = encoder_layerdrop snake_case = decoder_layerdrop snake_case = encoder_layers snake_case = auxiliary_loss snake_case = position_embedding_type snake_case = backbone snake_case = use_pretrained_backbone snake_case = dilation # Hungarian matcher snake_case = class_cost snake_case = bbox_cost snake_case = giou_cost # Loss coefficients snake_case = mask_loss_coefficient snake_case = dice_loss_coefficient snake_case = bbox_loss_coefficient snake_case = giou_loss_coefficient snake_case = eos_coefficient super().__init__(is_encoder_decoder=__snake_case , **__snake_case ) @property def a_ ( self ): return self.encoder_attention_heads @property def a_ ( self ): return self.d_model @classmethod def a_ ( cls , __snake_case , **__snake_case ): return cls(backbone_config=__snake_case , **__snake_case ) def a_ ( self ): snake_case = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: snake_case = self.backbone_config.to_dict() snake_case = self.__class__.model_type return output class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = version.parse('1.11' ) @property def a_ ( self ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def a_ ( self ): return 1E-5 @property def a_ ( self ): return 1_2
368
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor _SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) class A__ ( snake_case__ ): """simple docstring""" def __init__( self , *__snake_case , **__snake_case ): warnings.warn( '''The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use BeitImageProcessor instead.''' , __snake_case , ) super().__init__(*__snake_case , **__snake_case )
213
0
from __future__ import annotations def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case , __snake_case ) -> None: """simple docstring""" if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): _lowercase , _lowercase =array[indexa], array[indexa] def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case , __snake_case ) -> None: """simple docstring""" if length > 1: _lowercase =int(length / 2 ) for i in range(__snake_case , low + middle ): comp_and_swap(__snake_case , __snake_case , i + middle , __snake_case ) bitonic_merge(__snake_case , __snake_case , __snake_case , __snake_case ) bitonic_merge(__snake_case , low + middle , __snake_case , __snake_case ) def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case , __snake_case ) -> None: """simple docstring""" if length > 1: _lowercase =int(length / 2 ) bitonic_sort(__snake_case , __snake_case , __snake_case , 1 ) bitonic_sort(__snake_case , low + middle , __snake_case , 0 ) bitonic_merge(__snake_case , __snake_case , __snake_case , __snake_case ) if __name__ == "__main__": UpperCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase__ = [int(item.strip()) for item in user_input.split(''',''')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('''\nSorted array in ascending order is: ''', end='''''') print(*unsorted, sep=''', ''') bitonic_merge(unsorted, 0, len(unsorted), 0) print('''Sorted array in descending order is: ''', end='''''') print(*unsorted, sep=''', ''')
5
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : List[Any] = XGLMTokenizer _UpperCamelCase : List[Any] = XGLMTokenizerFast _UpperCamelCase : Dict = True _UpperCamelCase : Tuple = True def __A ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : List[Any] = XGLMTokenizer(a__ , keep_accents=a__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : List[str] = """<pad>""" _lowerCAmelCase : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a__ ) , a__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a__ ) , a__ ) def __A ( self ): _lowerCAmelCase : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(len(a__ ) , 1008 ) def __A ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def __A ( self ): _lowerCAmelCase : List[Any] = XGLMTokenizer(a__ , keep_accents=a__ ) _lowerCAmelCase : Dict = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : Any = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) _lowerCAmelCase : List[str] = tokenizer.convert_tokens_to_ids(a__ ) self.assertListEqual( a__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : Optional[int] = tokenizer.convert_ids_to_tokens(a__ ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) @cached_property def __A ( self ): return XGLMTokenizer.from_pretrained("""facebook/xglm-564M""" ) def __A ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(a__ , f.name ) _lowerCAmelCase : Union[str, Any] = XGLMTokenizer(f.name , keep_accents=a__ ) _lowerCAmelCase : List[str] = pickle.dumps(a__ ) pickle.loads(a__ ) def __A ( self ): if not self.test_rust_tokenizer: return _lowerCAmelCase : List[str] = self.get_tokenizer() _lowerCAmelCase : Optional[Any] = self.get_rust_tokenizer() _lowerCAmelCase : Tuple = """I was born in 92000, and this is falsé.""" _lowerCAmelCase : List[Any] = tokenizer.tokenize(a__ ) _lowerCAmelCase : Tuple = rust_tokenizer.tokenize(a__ ) self.assertListEqual(a__ , a__ ) _lowerCAmelCase : Union[str, Any] = tokenizer.encode(a__ , add_special_tokens=a__ ) _lowerCAmelCase : str = rust_tokenizer.encode(a__ , add_special_tokens=a__ ) self.assertListEqual(a__ , a__ ) _lowerCAmelCase : int = self.get_rust_tokenizer() _lowerCAmelCase : Dict = tokenizer.encode(a__ ) _lowerCAmelCase : List[Any] = rust_tokenizer.encode(a__ ) self.assertListEqual(a__ , a__ ) @slow def __A ( self ): _lowerCAmelCase : int = """Hello World!""" _lowerCAmelCase : Optional[int] = [2, 31227, 4447, 35] self.assertListEqual(a__ , self.big_tokenizer.encode(a__ ) ) @slow def __A ( self ): _lowerCAmelCase : Any = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth""" ) # fmt: off _lowerCAmelCase : List[str] = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 71630, 28085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 13675, 377, 652, 7580, 10341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 202277, 17892, 33, 60, 87, 4, 3234, 157, 61, 2667, 52376, 19, 88, 23, 735] # fmt: on self.assertListEqual(a__ , self.big_tokenizer.encode(a__ ) ) @slow def __A ( self ): # fmt: off _lowerCAmelCase : List[str] = { """input_ids""": [[2, 108825, 1163, 15, 88010, 473, 15898, 157, 13672, 1857, 312, 8, 238021, 1163, 53, 13672, 1857, 312, 8, 53283, 182396, 8, 18566, 16, 36733, 4101, 8, 230, 244017, 122553, 7, 15, 132597, 4, 293, 12511, 7610, 4, 3414, 132597, 9, 4, 32361, 362, 4, 734, 28512, 32569, 18, 4, 32361, 26096, 14982, 73, 18715, 21433, 235261, 15, 492, 12427, 16, 53, 18715, 21433, 65454, 15, 23659, 563, 16, 278, 597, 2843, 595, 7931, 182396, 64186, 22, 886, 595, 132981, 53, 25540, 3449, 43982, 39901, 5951, 878, 330, 4, 27694, 80269, 312, 53, 6517, 11780, 611, 20408, 5], [2, 6, 132597, 67, 42897, 33, 592, 8, 163729, 25540, 361, 136997, 109514, 173230, 7, 501, 60, 102913, 196, 5631, 235, 63243, 473, 6, 231757, 74, 5277, 7905, 53, 3095, 37317, 22, 454, 183874, 5], [2, 268, 31298, 46530, 6, 132935, 43831, 7, 597, 32, 24, 3688, 9865, 5]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a__ , model_name="""facebook/xglm-564M""" , padding=a__ , )
44
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: lowercase__ = None lowercase__ = logging.get_logger(__name__) lowercase__ = '▁' lowercase__ = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'}, 'tokenizer_file': { 'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json' }, } lowercase__ = { 'google/pegasus-xsum': 512, } class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = PegasusTokenizer a__ = ["""input_ids""", """attention_mask"""] def __init__( self , lowercase=None , lowercase=None , lowercase="<pad>" , lowercase="</s>" , lowercase="<unk>" , lowercase="<mask_2>" , lowercase="<mask_1>" , lowercase=None , lowercase=1_03 , **lowercase , ) -> List[Any]: '''simple docstring''' a__: Optional[Any] = offset if additional_special_tokens is not None: if not isinstance(lowercase , lowercase): raise TypeError( f'additional_special_tokens should be of type {type(lowercase)}, but is' f' {type(lowercase)}') a__: Union[str, Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(lowercase) , self.offset - 1) ] if len(set(lowercase)) != len(lowercase): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.') a__: str = additional_special_tokens_extended else: a__: Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset)] super().__init__( lowercase , tokenizer_file=lowercase , pad_token=lowercase , eos_token=lowercase , unk_token=lowercase , mask_token=lowercase , mask_token_sent=lowercase , offset=lowercase , additional_special_tokens=lowercase , **lowercase , ) a__: Optional[int] = vocab_file a__: Optional[int] = False if not self.vocab_file else True def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' a__: Any = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( 'There should be 3 special tokens: mask_token, pad_token, and eos_token +' f' {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}') return [1 if x in all_special_ids else 0 for x in seq] def lowerCamelCase_ ( self , lowercase , lowercase = None , lowercase = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(lowercase) elif token_ids_a is None: return self._special_token_mask(lowercase) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def lowerCamelCase_ ( self , lowercase , lowercase=None) -> List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowercase): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return a__: int = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase): copyfile(self.vocab_file , lowercase) return (out_vocab_file,)
359
"""simple docstring""" import argparse import math import traceback import dateutil.parser as date_parser import requests def __a ( _SCREAMING_SNAKE_CASE ) ->Tuple: a__: Tuple = {} a__: Tuple = job['started_at'] a__: int = job['completed_at'] a__: Any = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: Tuple = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: str = round((end_datetime - start_datetime).total_seconds() / 60.0 ) a__: Any = start a__: Dict = end a__: Optional[int] = duration_in_min return job_info def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: Tuple = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: int = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: str = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) return job_time except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') lowercase__ = parser.parse_args() lowercase__ = get_job_time(args.workflow_run_id) lowercase__ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"{k}: {v['duration']}")
203
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase : Optional[int] = { 'configuration_llama': ['LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LlamaConfig'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = ['LlamaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : int = ['LlamaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ 'LlamaForCausalLM', 'LlamaModel', 'LlamaPreTrainedModel', 'LlamaForSequenceClassification', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys __UpperCamelCase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
182
import math def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ = 10001 ): try: snake_case_ = int(SCREAMING_SNAKE_CASE__ ) except (TypeError, ValueError): raise TypeError('''Parameter nth must be int or castable to int.''' ) from None if nth <= 0: raise ValueError('''Parameter nth must be greater than or equal to one.''' ) snake_case_ = [] snake_case_ = 2 while len(SCREAMING_SNAKE_CASE__ ) < nth: if is_prime(SCREAMING_SNAKE_CASE__ ): primes.append(SCREAMING_SNAKE_CASE__ ) num += 1 else: num += 1 return primes[len(SCREAMING_SNAKE_CASE__ ) - 1] if __name__ == "__main__": print(f"""{solution() = }""")
8
0
'''simple docstring''' from __future__ import annotations from decimal import Decimal from numpy import array def __lowerCamelCase ( __snake_case : list[list[float]] ) -> list[list[float]]: """simple docstring""" A__ : List[Any] =Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(__snake_case ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix A__ : List[Any] =float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError("""This matrix has no inverse.""" ) # Creates a copy of the matrix with swapped positions of the elements A__ : Optional[Any] =[[0.0, 0.0], [0.0, 0.0]] A__ , A__ : List[Any] =matrix[1][1], matrix[0][0] A__ , A__ : Optional[Any] =-matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(__snake_case ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(__snake_case ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule A__ : Optional[Any] =float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError("""This matrix has no inverse.""" ) # Creating cofactor matrix A__ : List[str] =[ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] A__ : int =(d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) A__ : Any =-( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) A__ : str =(d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) A__ : Optional[Any] =-( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) A__ : Tuple =(d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) A__ : Tuple =-( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) A__ : str =(d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) A__ : List[Any] =-( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) A__ : Optional[int] =(d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) A__ : Dict =array(__snake_case ) for i in range(3 ): for j in range(3 ): A__ : Tuple =cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix A__ : Optional[Any] =array(__snake_case ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(__snake_case ) # Calculate the inverse of the matrix return [[float(d(__snake_case ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError("""Please provide a matrix of size 2x2 or 3x3.""" )
136
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient __snake_case : Optional[int] = WebClient(token=os.environ['CI_SLACK_BOT_TOKEN']) def __lowerCamelCase ( __snake_case : List[Any] ) -> Union[str, Any]: """simple docstring""" A__ : Any =test_results.split(""" """ ) A__ : List[Any] =0 A__ : Optional[int] =0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. A__ : Dict =expressions[-2] if """=""" in expressions[-1] else expressions[-1] for i, expression in enumerate(__snake_case ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def __lowerCamelCase ( __snake_case : str ) -> Optional[int]: """simple docstring""" A__ : Dict ={} A__ : List[Any] =None A__ : Any =False for line in failures_short_lines.split("""\n""" ): if re.search(r"""_ \[doctest\]""", __snake_case ): A__ : List[str] =True A__ : Optional[int] =line.split(""" """ )[2] elif in_error and not line.split(""" """ )[0].isdigit(): A__ : List[str] =line A__ : int =False return failures class lowerCamelCase : '''simple docstring''' def __init__( self : str , lowerCAmelCase_ : str , lowerCAmelCase_ : Dict ) -> Dict: '''simple docstring''' A__ : Any =title A__ : List[Any] =doc_test_results["""time_spent"""].split(""",""" )[0] A__ : str =doc_test_results["""success"""] A__ : str =doc_test_results["""failures"""] A__ : Optional[int] =self.n_success + self.n_failures # Failures and success of the modeling tests A__ : List[Any] =doc_test_results @property def lowercase__ ( self : Optional[int] ) -> str: '''simple docstring''' A__ : List[str] =[self._time_spent] A__ : str =0 for time in time_spent: A__ : Union[str, Any] =time.split(""":""" ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(lowerCAmelCase_ ) == 1: A__ : str =[0, 0, time_parts[0]] A__ , A__ , A__ : int =int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 36_00 + minutes * 60 + seconds A__ , A__ , A__ : Optional[int] =total_secs // 36_00, (total_secs % 36_00) // 60, total_secs % 60 return f"{int(lowerCAmelCase_ )}h{int(lowerCAmelCase_ )}m{int(lowerCAmelCase_ )}s" @property def lowercase__ ( self : Any ) -> Dict: '''simple docstring''' return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def lowercase__ ( self : Dict ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": f"🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.", "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def lowercase__ ( self : Tuple ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": ( f"There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in" f" {self.time}." ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def lowercase__ ( self : Optional[Any] ) -> Dict: '''simple docstring''' A__ : Optional[Any] =40 A__ : List[Any] ={k: v["""failed"""] for k, v in doc_test_results.items() if isinstance(lowerCAmelCase_ , lowerCAmelCase_ )} A__ : Union[str, Any] ="""""" for category, failures in category_failures.items(): if len(lowerCAmelCase_ ) == 0: continue if report != "": report += "\n\n" report += f"*{category} failures*:".ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(lowerCAmelCase_ ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f"The following examples had failures:\n\n\n{report}\n", }, } @property def lowercase__ ( self : Any ) -> str: '''simple docstring''' A__ : Tuple =[self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(lowerCAmelCase_ ) @staticmethod def lowercase__ ( ) -> Any: '''simple docstring''' A__ : Dict =[ { """type""": """section""", """text""": { """type""": """plain_text""", """text""": """There was an issue running the tests.""", }, """accessory""": { """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """Check Action results""", """emoji""": True}, """url""": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } ] print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(lowerCAmelCase_ )} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text="""There was an issue running the tests.""" , blocks=lowerCAmelCase_ , ) def lowercase__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(self.payload )} ) ) A__ : Tuple =f"{self.n_failures} failures out of {self.n_tests} tests," if self.n_failures else """All tests passed.""" A__ : Optional[int] =client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , blocks=self.payload , text=lowerCAmelCase_ , ) def lowercase__ ( self : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[Any] ) -> Any: '''simple docstring''' A__ : int ="""""" for key, value in failures.items(): A__ : Optional[int] =value[:2_00] + """ [Truncated]""" if len(lowerCAmelCase_ ) > 2_50 else value failures_text += f"*{key}*\n_{value}_\n\n" A__ : List[Any] =job_name A__ : Dict ={"""type""": """section""", """text""": {"""type""": """mrkdwn""", """text""": text}} if job_link is not None: A__ : Dict ={ """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """GitHub Action job""", """emoji""": True}, """url""": job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def lowercase__ ( self : str ) -> str: '''simple docstring''' if self.thread_ts is None: raise ValueError("""Can only post reply if a post has been made.""" ) A__ : Optional[Any] =self.doc_test_results.pop("""job_link""" ) self.doc_test_results.pop("""failures""" ) self.doc_test_results.pop("""success""" ) self.doc_test_results.pop("""time_spent""" ) A__ : Union[str, Any] =sorted(self.doc_test_results.items() , key=lambda lowerCAmelCase_ : t[0] ) for job, job_result in sorted_dict: if len(job_result["""failures"""] ): A__ : Optional[Any] =f"*Num failures* :{len(job_result['failed'] )} \n" A__ : Tuple =job_result["""failures"""] A__ : Optional[Any] =self.get_reply_blocks(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , text=lowerCAmelCase_ ) print("""Sending the following reply""" ) print(json.dumps({"""blocks""": blocks} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text=f"Results for {job}" , blocks=lowerCAmelCase_ , thread_ts=self.thread_ts["""ts"""] , ) time.sleep(1 ) def __lowerCamelCase ( ) -> Any: """simple docstring""" A__ : Optional[Any] =os.environ["""GITHUB_RUN_ID"""] A__ : Tuple =f"https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100" A__ : List[str] =requests.get(__snake_case ).json() A__ : List[str] ={} try: jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) A__ : Optional[int] =math.ceil((result["""total_count"""] - 100) / 100 ) for i in range(__snake_case ): A__ : List[Any] =requests.get(url + f"&page={i + 2}" ).json() jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) return jobs except Exception as e: print("""Unknown error, could not fetch links.""", __snake_case ) return {} def __lowerCamelCase ( __snake_case : str ) -> Union[str, Any]: """simple docstring""" A__ : Any ={} if os.path.exists(__snake_case ): A__ : str =os.listdir(__snake_case ) for file in files: try: with open(os.path.join(__snake_case, __snake_case ), encoding="""utf-8""" ) as f: A__ : Tuple =f.read() except UnicodeDecodeError as e: raise ValueError(f"Could not open {os.path.join(__snake_case, __snake_case )}." ) from e return _artifact def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" class lowerCamelCase : '''simple docstring''' def __init__( self : List[str] , lowerCAmelCase_ : str ) -> Any: '''simple docstring''' A__ : List[Any] =name A__ : str =[] def __str__( self : Optional[Any] ) -> List[Any]: '''simple docstring''' return self.name def lowercase__ ( self : Any , lowerCAmelCase_ : str ) -> Tuple: '''simple docstring''' self.paths.append({"""name""": self.name, """path""": path} ) A__ : Dict[str, Artifact] ={} A__ : int =filter(os.path.isdir, os.listdir() ) for directory in directories: A__ : List[Any] =directory if artifact_name not in _available_artifacts: A__ : str =Artifact(__snake_case ) _available_artifacts[artifact_name].add_path(__snake_case ) return _available_artifacts if __name__ == "__main__": __snake_case : List[str] = get_job_links() __snake_case : int = retrieve_available_artifacts() __snake_case : Union[str, Any] = collections.OrderedDict( [ ('*.py', 'API Examples'), ('*.md', 'MD Examples'), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' __snake_case : Dict = { v: { 'failed': [], 'failures': {}, } for v in docs.values() } # Link to the GitHub Action job __snake_case : List[Any] = github_actions_job_links.get('run_doctests') __snake_case : Tuple = available_artifacts['doc_tests_gpu_test_reports'].paths[0] __snake_case : Optional[int] = retrieve_artifact(artifact_path['name']) if "stats" in artifact: __snake_case , __snake_case , __snake_case : Optional[Any] = handle_test_results(artifact['stats']) __snake_case : Optional[Any] = failed __snake_case : Union[str, Any] = success __snake_case : Union[str, Any] = time_spent[1:-1] + ', ' __snake_case : int = extract_first_line_failure(artifact['failures_short']) for line in artifact["summary_short"].split('\n'): if re.search('FAILED', line): __snake_case : Optional[int] = line.replace('FAILED ', '') __snake_case : str = line.split()[0].replace('\n', '') if "::" in line: __snake_case , __snake_case : Optional[Any] = line.split('::') else: __snake_case , __snake_case : Any = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): __snake_case : List[Any] = docs[file_regex] doc_test_results[category]["failed"].append(test) __snake_case : List[str] = all_failures[test] if test in all_failures else 'N/A' __snake_case : Optional[Any] = failure break __snake_case : int = Message('🤗 Results of the doc tests.', doc_test_results) message.post() message.post_reply()
136
1
'''simple docstring''' from __future__ import annotations from decimal import Decimal from numpy import array def _UpperCamelCase ( __A ) -> list[list[float]]: '''simple docstring''' UpperCamelCase__ = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(__A ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix UpperCamelCase__ = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creates a copy of the matrix with swapped positions of the elements UpperCamelCase__ = [[0.0, 0.0], [0.0, 0.0]] UpperCamelCase__ , UpperCamelCase__ = matrix[1][1], matrix[0][0] UpperCamelCase__ , UpperCamelCase__ = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(__A ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(__A ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule UpperCamelCase__ = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creating cofactor matrix UpperCamelCase__ = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] UpperCamelCase__ = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) UpperCamelCase__ = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) UpperCamelCase__ = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) UpperCamelCase__ = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) UpperCamelCase__ = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) UpperCamelCase__ = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) UpperCamelCase__ = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) UpperCamelCase__ = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) UpperCamelCase__ = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) UpperCamelCase__ = array(__A ) for i in range(3 ): for j in range(3 ): UpperCamelCase__ = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix UpperCamelCase__ = array(__A ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(__A ) # Calculate the inverse of the matrix return [[float(d(__A ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError("Please provide a matrix of size 2x2 or 3x3." )
80
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Any = logging.get_logger(__name__) a__ : str = { 'SCUT-DLVCLab/lilt-roberta-en-base': ( 'https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json' ), } class lowercase_ ( a__ ): __UpperCAmelCase = 'lilt' def __init__( self , a=3_05_22 , a=7_68 , a=12 , a=12 , a=30_72 , a="gelu" , a=0.1 , a=0.1 , a=5_12 , a=2 , a=0.02 , a=1e-12 , a=0 , a="absolute" , a=None , a=4 , a=10_24 , **a , ): super().__init__(pad_token_id=a , **a ) UpperCamelCase__ = vocab_size UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = hidden_act UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = max_position_embeddings UpperCamelCase__ = type_vocab_size UpperCamelCase__ = initializer_range UpperCamelCase__ = layer_norm_eps UpperCamelCase__ = position_embedding_type UpperCamelCase__ = classifier_dropout UpperCamelCase__ = channel_shrink_ratio UpperCamelCase__ = max_ad_position_embeddings
80
1
"""simple docstring""" import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class _lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ ) def lowerCamelCase ( self ) -> Any: '''simple docstring''' snake_case : List[Any] = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(UpperCamelCase__ ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step , 3 ) self.assertEqual(len(accumulator.gradients ) , 1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [-2.0, 5.0] , tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [0.0, 0.0] , tol=1e-2 ) def lowerCamelCase ( self ) -> int: '''simple docstring''' snake_case : Optional[Any] = None ops.enable_eager_execution_internal() snake_case : str = tf.config.list_physical_devices("CPU" ) if len(UpperCamelCase__ ) == 1: tf.config.set_logical_device_configuration( physical_devices[0] , [tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) snake_case : Optional[int] = tf.config.list_logical_devices(device_type="CPU" ) snake_case : List[str] = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): snake_case : int = GradientAccumulator() snake_case : int = tf.Variable([4.0, 3.0] ) snake_case ,snake_case : Any = create_optimizer(5e-5 , 10 , 5 ) snake_case : Tuple = tf.Variable([0.0, 0.0] , trainable=UpperCamelCase__ ) def accumulate_on_replica(UpperCamelCase__ ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients , [variable] ) ) ) @tf.function def accumulate(UpperCamelCase__ , UpperCamelCase__ ): with strategy.scope(): snake_case : Union[str, Any] = strategy.experimental_local_results(UpperCamelCase__ ) local_variables[0].assign(UpperCamelCase__ ) local_variables[1].assign(UpperCamelCase__ ) strategy.run(UpperCamelCase__ , args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(UpperCamelCase__ ) def _check_local_values(UpperCamelCase__ , UpperCamelCase__ ): snake_case : List[Any] = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value() , UpperCamelCase__ , tol=1e-2 ) self.assertListAlmostEqual(values[1].value() , UpperCamelCase__ , tol=1e-2 ) accumulate([1.0, 2.0] , [-1.0, 1.0] ) accumulate([3.0, -1.0] , [-1.0, -1.0] ) accumulate([-2.0, 2.0] , [3.0, -2.0] ) self.assertEqual(accumulator.step , 3 ) _check_local_values([2.0, 3.0] , [1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value() , [4.0, 3.0] , tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) _check_local_values([0.0, 0.0] , [0.0, 0.0] )
112
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __snake_case = { """configuration_conditional_detr""": [ """CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConditionalDetrConfig""", """ConditionalDetrOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""ConditionalDetrFeatureExtractor"""] __snake_case = ["""ConditionalDetrImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ """CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConditionalDetrForObjectDetection""", """ConditionalDetrForSegmentation""", """ConditionalDetrModel""", """ConditionalDetrPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
112
1
import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase = AudioLDMPipeline __lowerCamelCase = TEXT_TO_AUDIO_PARAMS __lowerCamelCase = TEXT_TO_AUDIO_BATCH_PARAMS __lowerCamelCase = frozenset( [ '''num_inference_steps''', '''num_waveforms_per_prompt''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def snake_case ( self ): """simple docstring""" torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=(32, 64) , class_embed_type="""simple_projection""" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=_snake_case , ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=_snake_case , set_alpha_to_one=_snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , projection_dim=32 , ) _lowerCAmelCase = ClapTextModelWithProjection(_snake_case ) _lowerCAmelCase = RobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-roberta""" , model_max_length=77 ) _lowerCAmelCase = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=16000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=_snake_case , ) _lowerCAmelCase = SpeechTaHifiGan(_snake_case ) _lowerCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """vocoder""": vocoder, } return components def snake_case ( self , _snake_case , _snake_case=0 ): """simple docstring""" if str(_snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(_snake_case ) else: _lowerCAmelCase = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) _lowerCAmelCase = { """prompt""": """A hammer hitting a wooden surface""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, } return inputs def snake_case ( self ): """simple docstring""" _lowerCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = audioldm_pipe(**_snake_case ) _lowerCAmelCase = output.audios[0] assert audio.ndim == 1 assert len(_snake_case ) == 256 _lowerCAmelCase = audio[:10] _lowerCAmelCase = np.array( [-0.0050, 0.0050, -0.0060, 0.0033, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0033] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = 3 * [inputs["""prompt"""]] # forward _lowerCAmelCase = audioldm_pipe(**_snake_case ) _lowerCAmelCase = output.audios[0] _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = 3 * [inputs.pop("""prompt""" )] _lowerCAmelCase = audioldm_pipe.tokenizer( _snake_case , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_snake_case , return_tensors="""pt""" , ) _lowerCAmelCase = text_inputs["""input_ids"""].to(_snake_case ) _lowerCAmelCase = audioldm_pipe.text_encoder( _snake_case , ) _lowerCAmelCase = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state _lowerCAmelCase = F.normalize(_snake_case , dim=-1 ) _lowerCAmelCase = prompt_embeds # forward _lowerCAmelCase = audioldm_pipe(**_snake_case ) _lowerCAmelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = 3 * ["""this is a negative prompt"""] _lowerCAmelCase = negative_prompt _lowerCAmelCase = 3 * [inputs["""prompt"""]] # forward _lowerCAmelCase = audioldm_pipe(**_snake_case ) _lowerCAmelCase = output.audios[0] _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = 3 * [inputs.pop("""prompt""" )] _lowerCAmelCase = [] for p in [prompt, negative_prompt]: _lowerCAmelCase = audioldm_pipe.tokenizer( _snake_case , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_snake_case , return_tensors="""pt""" , ) _lowerCAmelCase = text_inputs["""input_ids"""].to(_snake_case ) _lowerCAmelCase = audioldm_pipe.text_encoder( _snake_case , ) _lowerCAmelCase = text_embeds.text_embeds # additional L_2 normalization over each hidden-state _lowerCAmelCase = F.normalize(_snake_case , dim=-1 ) embeds.append(_snake_case ) _lowerCAmelCase , _lowerCAmelCase = embeds # forward _lowerCAmelCase = audioldm_pipe(**_snake_case ) _lowerCAmelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = PNDMScheduler(skip_prk_steps=_snake_case ) _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = """egg cracking""" _lowerCAmelCase = audioldm_pipe(**_snake_case , negative_prompt=_snake_case ) _lowerCAmelCase = output.audios[0] assert audio.ndim == 1 assert len(_snake_case ) == 256 _lowerCAmelCase = audio[:10] _lowerCAmelCase = np.array( [-0.0051, 0.0050, -0.0060, 0.0034, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0032] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = PNDMScheduler(skip_prk_steps=_snake_case ) _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = """A hammer hitting a wooden surface""" # test num_waveforms_per_prompt=1 (default) _lowerCAmelCase = audioldm_pipe(_snake_case , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts _lowerCAmelCase = 2 _lowerCAmelCase = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt _lowerCAmelCase = 2 _lowerCAmelCase = audioldm_pipe(_snake_case , num_inference_steps=2 , num_waveforms_per_prompt=_snake_case ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts _lowerCAmelCase = 2 _lowerCAmelCase = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=_snake_case ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = audioldm_pipe.vocoder.config.sampling_rate _lowerCAmelCase = self.get_dummy_inputs(_snake_case ) _lowerCAmelCase = audioldm_pipe(audio_length_in_s=0.016 , **_snake_case ) _lowerCAmelCase = output.audios[0] assert audio.ndim == 1 assert len(_snake_case ) / vocoder_sampling_rate == 0.016 _lowerCAmelCase = audioldm_pipe(audio_length_in_s=0.032 , **_snake_case ) _lowerCAmelCase = output.audios[0] assert audio.ndim == 1 assert len(_snake_case ) / vocoder_sampling_rate == 0.032 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = AudioLDMPipeline(**_snake_case ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = ["""hey"""] _lowerCAmelCase = audioldm_pipe(_snake_case , num_inference_steps=1 ) _lowerCAmelCase = output.audios.shape assert audio_shape == (1, 256) _lowerCAmelCase = audioldm_pipe.vocoder.config config.model_in_dim *= 2 _lowerCAmelCase = SpeechTaHifiGan(_snake_case ).to(_snake_case ) _lowerCAmelCase = audioldm_pipe(_snake_case , num_inference_steps=1 ) _lowerCAmelCase = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def snake_case ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=_snake_case ) def snake_case ( self ): """simple docstring""" self._test_inference_batch_single_identical(test_mean_pixel_difference=_snake_case ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=_snake_case ) @slow class __lowerCAmelCase ( unittest.TestCase ): def snake_case ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case ( self , _snake_case , _snake_case="cpu" , _snake_case=torch.floataa , _snake_case=0 ): """simple docstring""" _lowerCAmelCase = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) _lowerCAmelCase = np.random.RandomState(_snake_case ).standard_normal((1, 8, 128, 16) ) _lowerCAmelCase = torch.from_numpy(_snake_case ).to(device=_snake_case , dtype=_snake_case ) _lowerCAmelCase = { """prompt""": """A hammer hitting a wooden surface""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 2.5, } return inputs def snake_case ( self ): """simple docstring""" _lowerCAmelCase = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_inputs(_snake_case ) _lowerCAmelCase = 25 _lowerCAmelCase = audioldm_pipe(**_snake_case ).audios[0] assert audio.ndim == 1 assert len(_snake_case ) == 81920 _lowerCAmelCase = audio[77230:77240] _lowerCAmelCase = np.array( [-0.4884, -0.4607, 0.0023, 0.5007, 0.5896, 0.5151, 0.3813, -0.0208, -0.3687, -0.4315] ) _lowerCAmelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def snake_case ( self ): """simple docstring""" _lowerCAmelCase = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) _lowerCAmelCase = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) _lowerCAmelCase = audioldm_pipe.to(_snake_case ) audioldm_pipe.set_progress_bar_config(disable=_snake_case ) _lowerCAmelCase = self.get_inputs(_snake_case ) _lowerCAmelCase = audioldm_pipe(**_snake_case ).audios[0] assert audio.ndim == 1 assert len(_snake_case ) == 81920 _lowerCAmelCase = audio[27780:27790] _lowerCAmelCase = np.array([-0.2131, -0.0873, -0.0124, -0.0189, 0.0569, 0.1373, 0.1883, 0.2886, 0.3297, 0.2212] ) _lowerCAmelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
82
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel __UpperCAmelCase : Any = { "text_branch": "text_model", "audio_branch": "audio_model.audio_encoder", "attn": "attention.self", "self.proj": "output.dense", "attention.self_mask": "attn_mask", "mlp.fc1": "intermediate.dense", "mlp.fc2": "output.dense", "norm1": "layernorm_before", "norm2": "layernorm_after", "bn0": "batch_norm", } __UpperCAmelCase : List[str] = AutoFeatureExtractor.from_pretrained("laion/clap-htsat-unfused", truncation="rand_trunc") def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False) -> Union[str, Any]: __snake_case , __snake_case: int = create_model( """HTSAT-tiny""" , """roberta""" , SCREAMING_SNAKE_CASE__ , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=SCREAMING_SNAKE_CASE__ , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def A__ ( SCREAMING_SNAKE_CASE__) -> Any: __snake_case: Optional[Any] = {} __snake_case: int = r""".*sequential.(\d+).*""" __snake_case: List[str] = r""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __snake_case: Tuple = key.replace(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) if re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): # replace sequential layers with list __snake_case: Optional[int] = re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__).group(1) __snake_case: str = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(SCREAMING_SNAKE_CASE__)//3}.linear.''') elif re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): __snake_case: Any = int(re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__).group(1)) # Because in CLAP they use `nn.Sequential`... __snake_case: Dict = 1 if projecton_layer == 0 else 2 __snake_case: Any = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''') if "audio" and "qkv" in key: # split qkv into query key and value __snake_case: List[str] = value __snake_case: Optional[Any] = mixed_qkv.size(0) // 3 __snake_case: Union[str, Any] = mixed_qkv[:qkv_dim] __snake_case: Dict = mixed_qkv[qkv_dim : qkv_dim * 2] __snake_case: int = mixed_qkv[qkv_dim * 2 :] __snake_case: Optional[Any] = query_layer __snake_case: str = key_layer __snake_case: int = value_layer else: __snake_case: Dict = value return model_state_dict def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False) -> Optional[Any]: __snake_case , __snake_case: List[str] = init_clap(SCREAMING_SNAKE_CASE__ , enable_fusion=SCREAMING_SNAKE_CASE__) clap_model.eval() __snake_case: List[str] = clap_model.state_dict() __snake_case: Optional[int] = rename_state_dict(SCREAMING_SNAKE_CASE__) __snake_case: Any = ClapConfig() __snake_case: Dict = enable_fusion __snake_case: List[str] = ClapModel(SCREAMING_SNAKE_CASE__) # ignore the spectrogram embedding layer model.load_state_dict(SCREAMING_SNAKE_CASE__ , strict=SCREAMING_SNAKE_CASE__) model.save_pretrained(SCREAMING_SNAKE_CASE__) transformers_config.save_pretrained(SCREAMING_SNAKE_CASE__) if __name__ == "__main__": __UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument("--enable_fusion", action="store_true", help="Whether to enable fusion or not") __UpperCAmelCase : Tuple = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
111
0
"""simple docstring""" class _a ( lowerCAmelCase): """simple docstring""" pass class _a ( lowerCAmelCase): """simple docstring""" pass class _a : """simple docstring""" def __init__( self : Tuple )->Optional[int]: _UpperCAmelCase = [ [], [], [], ] def lowercase__ ( self : List[str] , __UpperCamelCase : int , __UpperCamelCase : int )->None: try: if len(self.queues[priority] ) >= 1_0_0: raise OverflowError('''Maximum queue size is 100''' ) self.queues[priority].append(__UpperCamelCase ) except IndexError: raise ValueError('''Valid priorities are 0, 1, and 2''' ) def lowercase__ ( self : List[Any] )->int: for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError('''All queues are empty''' ) def __str__( self : Optional[Any] )->str: return "\n".join(F'Priority {i}: {q}' for i, q in enumerate(self.queues ) ) class _a : """simple docstring""" def __init__( self : Dict )->List[Any]: _UpperCAmelCase = [] def lowercase__ ( self : int , __UpperCamelCase : int )->None: if len(self.queue ) == 1_0_0: raise OverFlowError('''Maximum queue size is 100''' ) self.queue.append(__UpperCamelCase ) def lowercase__ ( self : str )->int: if not self.queue: raise UnderFlowError('''The queue is empty''' ) else: _UpperCAmelCase = min(self.queue ) self.queue.remove(__UpperCamelCase ) return data def __str__( self : str )->str: return str(self.queue ) def lowercase ( ): '''simple docstring''' _UpperCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def lowercase ( ): '''simple docstring''' _UpperCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
326
"""simple docstring""" from __future__ import annotations import numpy as np def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = np.shape(_SCREAMING_SNAKE_CASE ) if rows != columns: _UpperCAmelCase = ( '''\'table\' has to be of square shaped array but got a ''' f'{rows}x{columns} array:\n{table}' ) raise ValueError(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.zeros((rows, columns) ) _UpperCAmelCase = np.zeros((rows, columns) ) for i in range(_SCREAMING_SNAKE_CASE ): for j in range(_SCREAMING_SNAKE_CASE ): _UpperCAmelCase = sum(lower[i][k] * upper[k][j] for k in range(_SCREAMING_SNAKE_CASE ) ) if upper[j][j] == 0: raise ArithmeticError('''No LU decomposition exists''' ) _UpperCAmelCase = (table[i][j] - total) / upper[j][j] _UpperCAmelCase = 1 for j in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _UpperCAmelCase = sum(lower[i][k] * upper[k][j] for k in range(_SCREAMING_SNAKE_CASE ) ) _UpperCAmelCase = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
326
1
"""simple docstring""" from collections.abc import Callable import numpy as np def _snake_case ( UpperCamelCase : Callable , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float ): UpperCAmelCase : Any = int(np.ceil((x_end - xa) / step_size ) ) UpperCAmelCase : Optional[Any] = np.zeros((n + 1,) ) UpperCAmelCase : Optional[int] = ya UpperCAmelCase : int = xa for k in range(UpperCamelCase ): UpperCAmelCase : Optional[int] = y[k] + step_size * ode_func(UpperCamelCase , y[k] ) UpperCAmelCase : Optional[int] = y[k] + ( (step_size / 2) * (ode_func(UpperCamelCase , y[k] ) + ode_func(x + step_size , UpperCamelCase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
109
"""simple docstring""" import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase = logging.get_logger(__name__) def lowerCAmelCase_ ( snake_case_ : List[str] ) ->List[str]: print('Loading config file...' ) def flatten_yaml_as_dict(snake_case_ : str , snake_case_ : Optional[int]="" , snake_case_ : Optional[Any]="." ): lowerCamelCase__ : Union[str, Any] =[] for k, v in d.items(): lowerCamelCase__ : Any =parent_key + sep + k if parent_key else k if isinstance(snake_case_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(snake_case_ , snake_case_ , sep=snake_case_ ).items() ) else: items.append((new_key, v) ) return dict(snake_case_ ) lowerCamelCase__ : str =argparse.Namespace() with open(snake_case_ , 'r' ) as yaml_file: try: lowerCamelCase__ : Union[str, Any] =yaml.load(snake_case_ , Loader=yaml.FullLoader ) lowerCamelCase__ : List[Any] =flatten_yaml_as_dict(snake_case_ ) for k, v in flat_cfg.items(): setattr(snake_case_ , snake_case_ , snake_case_ ) except yaml.YAMLError as exc: logger.error('Error while loading config file: {}. Error message: {}'.format(snake_case_ , str(snake_case_ ) ) ) return config def lowerCAmelCase_ ( snake_case_ : Dict , snake_case_ : Optional[Any] ) ->List[Any]: lowerCamelCase__ : Tuple =MobileViTVaConfig() lowerCamelCase__ : List[Any] =False # dataset if task_name.startswith('imagenet1k_' ): lowerCamelCase__ : List[str] =1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: lowerCamelCase__ : Optional[int] =3_8_4 else: lowerCamelCase__ : Dict =2_5_6 lowerCamelCase__ : Dict ='imagenet-1k-id2label.json' elif task_name.startswith('imagenet21k_to_1k_' ): lowerCamelCase__ : Tuple =2_1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: lowerCamelCase__ : Tuple =3_8_4 else: lowerCamelCase__ : List[str] =2_5_6 lowerCamelCase__ : Any ='imagenet-22k-id2label.json' elif task_name.startswith('ade20k_' ): lowerCamelCase__ : int =1_5_1 lowerCamelCase__ : int =5_1_2 lowerCamelCase__ : Dict ='ade20k-id2label.json' lowerCamelCase__ : Union[str, Any] =True elif task_name.startswith('voc_' ): lowerCamelCase__ : Optional[Any] =2_1 lowerCamelCase__ : Dict =5_1_2 lowerCamelCase__ : Union[str, Any] ='pascal-voc-id2label.json' lowerCamelCase__ : int =True # orig_config lowerCamelCase__ : Union[str, Any] =load_orig_config_file(snake_case_ ) assert getattr(snake_case_ , 'model.classification.name' , -1 ) == "mobilevit_v2", "Invalid model" lowerCamelCase__ : List[str] =getattr(snake_case_ , 'model.classification.mitv2.width_multiplier' , 1.0 ) assert ( getattr(snake_case_ , 'model.classification.mitv2.attn_norm_layer' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" lowerCamelCase__ : List[Any] =getattr(snake_case_ , 'model.classification.activation.name' , 'swish' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: lowerCamelCase__ : str =getattr(snake_case_ , 'model.segmentation.output_stride' , 1_6 ) if "_deeplabv3" in task_name: lowerCamelCase__ : Dict =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_rates' , [1_2, 2_4, 3_6] ) lowerCamelCase__ : str =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_out_channels' , 5_1_2 ) lowerCamelCase__ : Dict =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_dropout' , 0.1 ) # id2label lowerCamelCase__ : Union[str, Any] ='huggingface/label-files' lowerCamelCase__ : List[Any] =json.load(open(hf_hub_download(snake_case_ , snake_case_ , repo_type='dataset' ) , 'r' ) ) lowerCamelCase__ : List[str] ={int(snake_case_ ): v for k, v in idalabel.items()} lowerCamelCase__ : Optional[int] =idalabel lowerCamelCase__ : List[str] ={v: k for k, v in idalabel.items()} return config def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : List[str] , snake_case_ : Tuple ) ->Tuple: lowerCamelCase__ : List[Any] =dct.pop(snake_case_ ) lowerCamelCase__ : str =val def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Any=False ) ->str: if base_model: lowerCamelCase__ : Tuple ='' else: lowerCamelCase__ : Optional[Any] ='mobilevitv2.' lowerCamelCase__ : Tuple =[] for k in state_dict.keys(): if k[:8] == "encoder.": lowerCamelCase__ : Tuple =k[8:] else: lowerCamelCase__ : Tuple =k if ".block." in k: lowerCamelCase__ : Optional[Any] =k_new.replace('.block.' , '.' ) if ".conv." in k: lowerCamelCase__ : Union[str, Any] =k_new.replace('.conv.' , '.convolution.' ) if ".norm." in k: lowerCamelCase__ : Dict =k_new.replace('.norm.' , '.normalization.' ) if "conv_1." in k: lowerCamelCase__ : Dict =k_new.replace('conv_1.' , f"""{model_prefix}conv_stem.""" ) for i in [1, 2]: if f"""layer_{i}.""" in k: lowerCamelCase__ : Union[str, Any] =k_new.replace(f"""layer_{i}.""" , f"""{model_prefix}encoder.layer.{i-1}.layer.""" ) if ".exp_1x1." in k: lowerCamelCase__ : Tuple =k_new.replace('.exp_1x1.' , '.expand_1x1.' ) if ".red_1x1." in k: lowerCamelCase__ : Union[str, Any] =k_new.replace('.red_1x1.' , '.reduce_1x1.' ) for i in [3, 4, 5]: if f"""layer_{i}.0.""" in k: lowerCamelCase__ : Tuple =k_new.replace(f"""layer_{i}.0.""" , f"""{model_prefix}encoder.layer.{i-1}.downsampling_layer.""" ) if f"""layer_{i}.1.local_rep.0.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.local_rep.0.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_kxk.""" ) if f"""layer_{i}.1.local_rep.1.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.local_rep.1.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_1x1.""" ) for i in [3, 4, 5]: if i == 3: lowerCamelCase__ : Optional[int] =[0, 1] elif i == 4: lowerCamelCase__ : int =[0, 1, 2, 3] elif i == 5: lowerCamelCase__ : List[str] =[0, 1, 2] for j in j_in: if f"""layer_{i}.1.global_rep.{j}.""" in k: lowerCamelCase__ : Any =k_new.replace( f"""layer_{i}.1.global_rep.{j}.""" , f"""{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.""" ) if f"""layer_{i}.1.global_rep.{j+1}.""" in k: lowerCamelCase__ : Optional[Any] =k_new.replace( f"""layer_{i}.1.global_rep.{j+1}.""" , f"""{model_prefix}encoder.layer.{i-1}.layernorm.""" ) if f"""layer_{i}.1.conv_proj.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.conv_proj.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_projection.""" ) if "pre_norm_attn.0." in k: lowerCamelCase__ : int =k_new.replace('pre_norm_attn.0.' , 'layernorm_before.' ) if "pre_norm_attn.1." in k: lowerCamelCase__ : Optional[int] =k_new.replace('pre_norm_attn.1.' , 'attention.' ) if "pre_norm_ffn.0." in k: lowerCamelCase__ : List[str] =k_new.replace('pre_norm_ffn.0.' , 'layernorm_after.' ) if "pre_norm_ffn.1." in k: lowerCamelCase__ : Any =k_new.replace('pre_norm_ffn.1.' , 'ffn.conv1.' ) if "pre_norm_ffn.3." in k: lowerCamelCase__ : Tuple =k_new.replace('pre_norm_ffn.3.' , 'ffn.conv2.' ) if "classifier.1." in k: lowerCamelCase__ : List[Any] =k_new.replace('classifier.1.' , 'classifier.' ) if "seg_head." in k: lowerCamelCase__ : int =k_new.replace('seg_head.' , 'segmentation_head.' ) if ".aspp_layer." in k: lowerCamelCase__ : List[str] =k_new.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in k: lowerCamelCase__ : Optional[int] =k_new.replace('.aspp_pool.' , '.' ) rename_keys.append((k, k_new) ) return rename_keys def lowerCAmelCase_ ( snake_case_ : Tuple ) ->str: lowerCamelCase__ : str =[] for k in state_dict.keys(): if k.startswith('seg_head.aux_head.' ): keys_to_ignore.append(snake_case_ ) for k in keys_to_ignore: state_dict.pop(snake_case_ , snake_case_ ) def lowerCAmelCase_ ( ) ->Optional[int]: lowerCamelCase__ : List[str] ='http://images.cocodataset.org/val2017/000000039769.jpg' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" lowerCamelCase__ : Any =Image.open(requests.get(snake_case_ , stream=snake_case_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Dict , snake_case_ : Any , snake_case_ : List[str] ) ->Optional[int]: lowerCamelCase__ : int =get_mobilevitva_config(snake_case_ , snake_case_ ) # load original state_dict lowerCamelCase__ : List[str] =torch.load(snake_case_ , map_location='cpu' ) # load huggingface model if task_name.startswith('ade20k_' ) or task_name.startswith('voc_' ): lowerCamelCase__ : List[str] =MobileViTVaForSemanticSegmentation(snake_case_ ).eval() lowerCamelCase__ : Optional[Any] =False else: lowerCamelCase__ : Union[str, Any] =MobileViTVaForImageClassification(snake_case_ ).eval() lowerCamelCase__ : List[Any] =False # remove and rename some keys of load the original model lowerCamelCase__ : Union[str, Any] =checkpoint remove_unused_keys(snake_case_ ) lowerCamelCase__ : List[str] =create_rename_keys(snake_case_ , base_model=snake_case_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(snake_case_ , snake_case_ , snake_case_ ) # load modified state_dict model.load_state_dict(snake_case_ ) # Check outputs on an image, prepared by MobileViTImageProcessor lowerCamelCase__ : str =MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 3_2 ) lowerCamelCase__ : int =image_processor(images=prepare_img() , return_tensors='pt' ) lowerCamelCase__ : Optional[int] =model(**snake_case_ ) # verify classification model if task_name.startswith('imagenet' ): lowerCamelCase__ : Tuple =outputs.logits lowerCamelCase__ : Optional[int] =logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('imagenet1k_256' ) and config.width_multiplier == 1.0: # expected_logits for base variant lowerCamelCase__ : Tuple =torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ) assert torch.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) Path(snake_case_ ).mkdir(exist_ok=snake_case_ ) print(f"""Saving model {task_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(snake_case_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(snake_case_ ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you'd like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCAmelCase = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
126
0
"""simple docstring""" from math import ceil, sqrt def UpperCAmelCase__ (snake_case__ : int = 1_00_00_00 ): """simple docstring""" _snake_case : Dict = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: _snake_case : List[Any] = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: _snake_case : Any = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F'''{solution() = }''')
360
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 1_00_00_00 ): """simple docstring""" _snake_case : Dict = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , snake_case__ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
132
0
import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class lowercase__ ( unittest.TestCase): UpperCamelCase_ = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING UpperCamelCase_ = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def __A ( self : Dict , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = TextaTextGenerationPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) return generator, ["Something to write", "Something else"] def __A ( self : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = generator('''Something there''' ) self.assertEqual(UpperCamelCase__ , [{'''generated_text''': ANY(UpperCamelCase__ )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['''generated_text'''].startswith('''Something there''' ) ) SCREAMING_SNAKE_CASE : Any = generator(['''This is great !''', '''Something else'''] , num_return_sequences=2 , do_sample=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__ , [ [{'''generated_text''': ANY(UpperCamelCase__ )}, {'''generated_text''': ANY(UpperCamelCase__ )}], [{'''generated_text''': ANY(UpperCamelCase__ )}, {'''generated_text''': ANY(UpperCamelCase__ )}], ] , ) SCREAMING_SNAKE_CASE : Union[str, Any] = generator( ['''This is great !''', '''Something else'''] , num_return_sequences=2 , batch_size=2 , do_sample=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__ , [ [{'''generated_text''': ANY(UpperCamelCase__ )}, {'''generated_text''': ANY(UpperCamelCase__ )}], [{'''generated_text''': ANY(UpperCamelCase__ )}, {'''generated_text''': ANY(UpperCamelCase__ )}], ] , ) with self.assertRaises(UpperCamelCase__ ): generator(4 ) @require_torch def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''pt''' ) # do_sample=False necessary for reproducibility SCREAMING_SNAKE_CASE : Dict = generator('''Something there''' , do_sample=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , [{'''generated_text''': ''''''}] ) SCREAMING_SNAKE_CASE : str = 3 SCREAMING_SNAKE_CASE : Optional[int] = generator( '''Something there''' , num_return_sequences=UpperCamelCase__ , num_beams=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : int = [ {'''generated_text''': '''Beide Beide Beide Beide Beide Beide Beide Beide Beide'''}, {'''generated_text''': '''Beide Beide Beide Beide Beide Beide Beide Beide'''}, {'''generated_text''': ''''''}, ] self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = generator('''This is a test''' , do_sample=UpperCamelCase__ , num_return_sequences=2 , return_tensors=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__ , [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ] , ) SCREAMING_SNAKE_CASE : str = generator.model.config.eos_token_id SCREAMING_SNAKE_CASE : Optional[Any] = '''<pad>''' SCREAMING_SNAKE_CASE : Tuple = generator( ['''This is a test''', '''This is a second test'''] , do_sample=UpperCamelCase__ , num_return_sequences=2 , batch_size=2 , return_tensors=UpperCamelCase__ , ) self.assertEqual( UpperCamelCase__ , [ [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], ] , ) @require_tf def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''tf''' ) # do_sample=False necessary for reproducibility SCREAMING_SNAKE_CASE : Optional[Any] = generator('''Something there''' , do_sample=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , [{'''generated_text''': ''''''}] )
182
'''simple docstring''' import logging import os from .state import PartialState class a__ ( logging.LoggerAdapter ): @staticmethod def SCREAMING_SNAKE_CASE__ ( a : Optional[Any] ): """simple docstring""" __lowerCamelCase = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def SCREAMING_SNAKE_CASE__ ( self : int , a : Optional[int] , a : str , *a : Optional[int] , **a : List[Any] ): """simple docstring""" if PartialState._shared_state == {}: raise RuntimeError( '''You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.''' ) __lowerCamelCase = kwargs.pop('''main_process_only''' , a ) __lowerCamelCase = kwargs.pop('''in_order''' , a ) if self.isEnabledFor(a ): if self._should_log(a ): __lowerCamelCase , __lowerCamelCase = self.process(a , a ) self.logger.log(a , a , *a , **a ) elif in_order: __lowerCamelCase = PartialState() for i in range(state.num_processes ): if i == state.process_index: __lowerCamelCase , __lowerCamelCase = self.process(a , a ) self.logger.log(a , a , *a , **a ) state.wait_for_everyone() def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ = None ) -> Optional[int]: if log_level is None: __lowerCamelCase = os.environ.get('''ACCELERATE_LOG_LEVEL''' , UpperCamelCase__ ) __lowerCamelCase = logging.getLogger(UpperCamelCase__ ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(UpperCamelCase__ , {} )
67
0
'''simple docstring''' def UpperCamelCase ( a ) -> Dict: '''simple docstring''' __magic_name__ = len(A__ ) for i in range(length - 1 ): __magic_name__ = i for k in range(i + 1 , A__ ): if collection[k] < collection[least]: __magic_name__ = k if least != i: __magic_name__ , __magic_name__ = (collection[i], collection[least]) return collection if __name__ == "__main__": _lowerCAmelCase = input("Enter numbers separated by a comma:\n").strip() _lowerCAmelCase = [int(item) for item in user_input.split(",")] print(selection_sort(unsorted))
358
'''simple docstring''' import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase = logging.get_logger(__name__) def UpperCamelCase ( a ) -> str: '''simple docstring''' __magic_name__ = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) __magic_name__ = MaskFormerConfig(backbone_config=a ) __magic_name__ = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok __magic_name__ = 847 __magic_name__ = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok __magic_name__ = 150 __magic_name__ = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok __magic_name__ = 171 __magic_name__ = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO __magic_name__ = 133 __magic_name__ = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok __magic_name__ = 19 __magic_name__ = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok __magic_name__ = 65 __magic_name__ = '''mapillary-vistas-id2label.json''' __magic_name__ = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) __magic_name__ = {int(a ): v for k, v in idalabel.items()} return config def UpperCamelCase ( a ) -> Tuple: '''simple docstring''' __magic_name__ = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_index''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((F'''backbone.layers.{i}.downsample.reduction.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((F'''backbone.norm{i}.weight''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.weight''') ) rename_keys.append((F'''backbone.norm{i}.bias''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.bias''') ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((F'''sem_seg_head.adapter_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias''') ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias''') ) # cross-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias''') ) # MLP 1 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc1.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc1.bias''') ) # MLP 2 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc2.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc2.bias''') ) # layernorm 1 (self-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias''') ) # layernorm 2 (cross-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias''') ) # layernorm 3 (final layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.weight''', F'''mask_embedder.{i}.0.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.bias''', F'''mask_embedder.{i}.0.bias''') ) # fmt: on return rename_keys def UpperCamelCase ( a , a , a ) -> str: '''simple docstring''' __magic_name__ = dct.pop(a ) __magic_name__ = val def UpperCamelCase ( a , a ) -> List[str]: '''simple docstring''' __magic_name__ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __magic_name__ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.weight''' ) __magic_name__ = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[:dim, :] __magic_name__ = in_proj_bias[: dim] __magic_name__ = in_proj_weight[ dim : dim * 2, : ] __magic_name__ = in_proj_bias[ dim : dim * 2 ] __magic_name__ = in_proj_weight[ -dim :, : ] __magic_name__ = in_proj_bias[-dim :] # fmt: on def UpperCamelCase ( a , a ) -> int: '''simple docstring''' # fmt: off __magic_name__ = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight''' ) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[: hidden_size, :] __magic_name__ = in_proj_bias[:config.hidden_size] __magic_name__ = in_proj_weight[hidden_size : hidden_size * 2, :] __magic_name__ = in_proj_bias[hidden_size : hidden_size * 2] __magic_name__ = in_proj_weight[-hidden_size :, :] __magic_name__ = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight''' ) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[: hidden_size, :] __magic_name__ = in_proj_bias[:config.hidden_size] __magic_name__ = in_proj_weight[hidden_size : hidden_size * 2, :] __magic_name__ = in_proj_bias[hidden_size : hidden_size * 2] __magic_name__ = in_proj_weight[-hidden_size :, :] __magic_name__ = in_proj_bias[-hidden_size :] # fmt: on def UpperCamelCase ( ) -> torch.Tensor: '''simple docstring''' __magic_name__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __magic_name__ = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def UpperCamelCase ( a , a , a , a = False ) -> Dict: '''simple docstring''' __magic_name__ = get_maskformer_config(a ) # load original state_dict with open(a , '''rb''' ) as f: __magic_name__ = pickle.load(a ) __magic_name__ = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys __magic_name__ = create_rename_keys(a ) for src, dest in rename_keys: rename_key(a , a , a ) read_in_swin_q_k_v(a , config.backbone_config ) read_in_decoder_q_k_v(a , a ) # update to torch tensors for key, value in state_dict.items(): __magic_name__ = torch.from_numpy(a ) # load 🤗 model __magic_name__ = MaskFormerForInstanceSegmentation(a ) model.eval() for name, param in model.named_parameters(): print(a , param.shape ) __magic_name__ , __magic_name__ = model.load_state_dict(a , strict=a ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(a ) == 0, F'''Unexpected keys: {unexpected_keys}''' # verify results __magic_name__ = prepare_img() if "vistas" in model_name: __magic_name__ = 65 elif "cityscapes" in model_name: __magic_name__ = 6_5535 else: __magic_name__ = 255 __magic_name__ = True if '''ade''' in model_name else False __magic_name__ = MaskFormerImageProcessor(ignore_index=a , reduce_labels=a ) __magic_name__ = image_processor(a , return_tensors='''pt''' ) __magic_name__ = model(**a ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": __magic_name__ = torch.tensor( [[3.63_53, -4.47_70, -2.60_65], [0.50_81, -4.23_94, -3.53_43], [2.19_09, -5.03_53, -1.93_23]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , a , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and image processor to {pytorch_dump_folder_path}''' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(F'''nielsr/{model_name}''' ) image_processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="maskformer-swin-tiny-ade", type=str, help=("Name of the MaskFormer model you'd like to convert",), ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl", type=str, help="Path to the original state dict (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _lowerCAmelCase = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
98
0
'''simple docstring''' import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() a__ : int = logging.get_logger(__name__) def _UpperCamelCase ( __A , __A , __A ) -> List[str]: '''simple docstring''' UpperCamelCase__ = UniSpeechSatForSequenceClassification.from_pretrained(__A , config=__A ) UpperCamelCase__ = downstream_dict["projector.weight"] UpperCamelCase__ = downstream_dict["projector.bias"] UpperCamelCase__ = downstream_dict["model.post_net.linear.weight"] UpperCamelCase__ = downstream_dict["model.post_net.linear.bias"] return model def _UpperCamelCase ( __A , __A , __A ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ = UniSpeechSatForAudioFrameClassification.from_pretrained(__A , config=__A ) UpperCamelCase__ = downstream_dict["model.linear.weight"] UpperCamelCase__ = downstream_dict["model.linear.bias"] return model def _UpperCamelCase ( __A , __A , __A ) -> str: '''simple docstring''' UpperCamelCase__ = UniSpeechSatForXVector.from_pretrained(__A , config=__A ) UpperCamelCase__ = downstream_dict["connector.weight"] UpperCamelCase__ = downstream_dict["connector.bias"] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): UpperCamelCase__ = downstream_dict[ F'''model.framelevel_feature_extractor.module.{i}.kernel.weight''' ] UpperCamelCase__ = downstream_dict[F'''model.framelevel_feature_extractor.module.{i}.kernel.bias'''] UpperCamelCase__ = downstream_dict["model.utterancelevel_feature_extractor.linear1.weight"] UpperCamelCase__ = downstream_dict["model.utterancelevel_feature_extractor.linear1.bias"] UpperCamelCase__ = downstream_dict["model.utterancelevel_feature_extractor.linear2.weight"] UpperCamelCase__ = downstream_dict["model.utterancelevel_feature_extractor.linear2.bias"] UpperCamelCase__ = downstream_dict["objective.W"] return model @torch.no_grad() def _UpperCamelCase ( __A , __A , __A , __A ) -> List[str]: '''simple docstring''' UpperCamelCase__ = torch.load(__A , map_location="cpu" ) UpperCamelCase__ = checkpoint["Downstream"] UpperCamelCase__ = UniSpeechSatConfig.from_pretrained(__A ) UpperCamelCase__ = WavaVecaFeatureExtractor.from_pretrained( __A , return_attention_mask=__A , do_normalize=__A ) UpperCamelCase__ = hf_config.architectures[0] if arch.endswith("ForSequenceClassification" ): UpperCamelCase__ = convert_classification(__A , __A , __A ) elif arch.endswith("ForAudioFrameClassification" ): UpperCamelCase__ = convert_diarization(__A , __A , __A ) elif arch.endswith("ForXVector" ): UpperCamelCase__ = convert_xvector(__A , __A , __A ) else: raise NotImplementedError(F'''S3PRL weights conversion is not supported for {arch}''' ) if hf_config.use_weighted_layer_sum: UpperCamelCase__ = checkpoint["Featurizer"]["weights"] hf_feature_extractor.save_pretrained(__A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": a__ : Tuple = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') a__ : Union[str, Any] = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
80
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ : Union[str, Any] = {'configuration_mbart': ['MBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MBartConfig', 'MBartOnnxConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = ['MBartTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[Any] = ['MBartTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[str] = [ 'MBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'MBartForCausalLM', 'MBartForConditionalGeneration', 'MBartForQuestionAnswering', 'MBartForSequenceClassification', 'MBartModel', 'MBartPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[str] = [ 'TFMBartForConditionalGeneration', 'TFMBartModel', 'TFMBartPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str = [ 'FlaxMBartForConditionalGeneration', 'FlaxMBartForQuestionAnswering', 'FlaxMBartForSequenceClassification', 'FlaxMBartModel', 'FlaxMBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys a__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
80
1
from __future__ import annotations import math def __UpperCamelCase ( _A : float , _A : int ) ->float: """simple docstring""" lowerCamelCase_ =u for i in range(1 , _A ): lowerCamelCase_ =temp * (u - i) return temp def __UpperCamelCase ( ) ->None: """simple docstring""" lowerCamelCase_ =int(input("""enter the numbers of values: """ ) ) lowerCamelCase_ =[] for _ in range(_A ): y.append([] ) for i in range(_A ): for j in range(_A ): y[i].append(_A ) lowerCamelCase_ =0 print("""enter the values of parameters in a list: """ ) lowerCamelCase_ =list(map(_A , input().split() ) ) print("""enter the values of corresponding parameters: """ ) for i in range(_A ): lowerCamelCase_ =float(input() ) lowerCamelCase_ =int(input("""enter the value to interpolate: """ ) ) lowerCamelCase_ =(value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , _A ): for j in range(n - i ): lowerCamelCase_ =y[j + 1][i - 1] - y[j][i - 1] lowerCamelCase_ =y[0][0] for i in range(1 , _A ): summ += (ucal(_A , _A ) * y[0][i]) / math.factorial(_A ) print(f'the value at {value} is {summ}' ) if __name__ == "__main__": main()
49
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def _snake_case ( self )-> Union[str, Any]: lowerCamelCase_ =SMALL_MODEL_IDENTIFIER lowerCamelCase_ ="""pt""" lowerCamelCase_ ="""tf""" def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> List[str]: lowerCamelCase_ =AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Optional[Any]: lowerCamelCase_ =TFAutoModel.from_pretrained(self.test_model , from_pt=_SCREAMING_SNAKE_CASE ) model_tf.save_pretrained(_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ ="""mock_framework""" # Framework provided - return whatever the user provides lowerCamelCase_ =FeaturesManager.determine_framework(self.test_model , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =FeaturesManager.determine_framework(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =FeaturesManager.determine_framework(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> Tuple: # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =FeaturesManager.determine_framework(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =FeaturesManager.determine_framework(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_SCREAMING_SNAKE_CASE ): lowerCamelCase_ =FeaturesManager.determine_framework(_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) with patch("""transformers.onnx.features.is_tf_available""" , _SCREAMING_SNAKE_CASE ): lowerCamelCase_ =FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_SCREAMING_SNAKE_CASE , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) with patch("""transformers.onnx.features.is_torch_available""" , _SCREAMING_SNAKE_CASE ): lowerCamelCase_ =FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_SCREAMING_SNAKE_CASE , self.framework_tf ) # Both in environment -> use PyTorch lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) with patch("""transformers.onnx.features.is_tf_available""" , _SCREAMING_SNAKE_CASE ), patch( """transformers.onnx.features.is_torch_available""" , _SCREAMING_SNAKE_CASE ): lowerCamelCase_ =FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_SCREAMING_SNAKE_CASE , self.framework_pt ) # Both not in environment -> raise error lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =MagicMock(return_value=_SCREAMING_SNAKE_CASE ) with patch("""transformers.onnx.features.is_tf_available""" , _SCREAMING_SNAKE_CASE ), patch( """transformers.onnx.features.is_torch_available""" , _SCREAMING_SNAKE_CASE ): with self.assertRaises(_SCREAMING_SNAKE_CASE ): lowerCamelCase_ =FeaturesManager.determine_framework(self.test_model )
49
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer lowerCamelCase__ : Any = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast lowerCamelCase__ : List[str] = TaTokenizerFast lowerCamelCase__ : Union[str, Any] = {'configuration_mt5': ['MT5Config', 'MT5OnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[int] = [ 'MT5EncoderModel', 'MT5ForConditionalGeneration', 'MT5ForQuestionAnswering', 'MT5Model', 'MT5PreTrainedModel', 'MT5Stack', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[int] = ['TFMT5EncoderModel', 'TFMT5ForConditionalGeneration', 'TFMT5Model'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[Any] = ['FlaxMT5EncoderModel', 'FlaxMT5ForConditionalGeneration', 'FlaxMT5Model'] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys lowerCamelCase__ : Optional[int] = _LazyModule( __name__, globals()['__file__'], _import_structure, extra_objects={'MT5Tokenizer': MTaTokenizer, 'MT5TokenizerFast': MTaTokenizerFast}, module_spec=__spec__, )
225
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCamelCase_ : '''simple docstring''' def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : Tuple=13 , _lowerCAmelCase : Dict=30 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Dict=3 , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=32 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Tuple=4 , _lowerCAmelCase : Optional[int]=37 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : Optional[Any]=0.1 , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : Optional[Any]=10 , _lowerCAmelCase : Dict=0.02 , _lowerCAmelCase : Optional[Any]=3 , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : List[Any]=2 , ): SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = image_size SCREAMING_SNAKE_CASE_ = patch_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = type_sequence_label_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = scope SCREAMING_SNAKE_CASE_ = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) SCREAMING_SNAKE_CASE_ = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE_ = num_patches + 2 def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = None if self.use_labels: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, pixel_values, labels def lowerCAmelCase_ ( self : Tuple ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def lowerCAmelCase_ ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any ): SCREAMING_SNAKE_CASE_ = TFDeiTModel(config=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ): SCREAMING_SNAKE_CASE_ = TFDeiTForMaskedImageModeling(config=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = TFDeiTForMaskedImageModeling(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def lowerCAmelCase_ ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ): SCREAMING_SNAKE_CASE_ = self.type_sequence_label_size SCREAMING_SNAKE_CASE_ = TFDeiTForImageClassification(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = TFDeiTForImageClassification(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = config_and_inputs SCREAMING_SNAKE_CASE_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) lowercase_ = ( { "feature-extraction": TFDeiTModel, "image-classification": (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) lowercase_ = False lowercase_ = False lowercase_ = False lowercase_ = False def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = TFDeiTModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def lowerCAmelCase_ ( self : List[str] ): self.config_tester.run_common_tests() @unittest.skip(reason='DeiT does not use inputs_embeds' ) def lowerCAmelCase_ ( self : Any ): pass def lowerCAmelCase_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) SCREAMING_SNAKE_CASE_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Dense ) ) def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any]=False ): SCREAMING_SNAKE_CASE_ = super()._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def lowerCAmelCase_ ( self : Union[str, Any] ): for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ = TFDeiTModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def UpperCAmelCase_ ( ) -> str: SCREAMING_SNAKE_CASE_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self : List[str] ): return ( DeiTImageProcessor.from_pretrained('facebook/deit-base-distilled-patch16-224' ) if is_vision_available() else None ) @slow def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = TFDeiTForImageClassificationWithTeacher.from_pretrained('facebook/deit-base-distilled-patch16-224' ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=_lowerCAmelCase , return_tensors='tf' ) # forward pass SCREAMING_SNAKE_CASE_ = model(**_lowerCAmelCase ) # verify the logits SCREAMING_SNAKE_CASE_ = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tf.constant([-1.0266, 0.1912, -1.2861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
225
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _snake_case : lowerCAmelCase :str = PegasusConfig lowerCAmelCase :Optional[int] = {} lowerCAmelCase :Dict = '''gelu''' def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=False , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=2 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=40 , _lowerCamelCase=2 , _lowerCamelCase=1 , _lowerCamelCase=0 , ): UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : Union[str, Any] = seq_length UpperCAmelCase__ : str = is_training UpperCAmelCase__ : Dict = use_labels UpperCAmelCase__ : Optional[int] = vocab_size UpperCAmelCase__ : int = hidden_size UpperCAmelCase__ : Union[str, Any] = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : Optional[Any] = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Any = pad_token_id UpperCAmelCase__ : Dict = bos_token_id def snake_case__ ( self): UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) UpperCAmelCase__ : str = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) UpperCAmelCase__ : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1) UpperCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase__ : Any = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase__ : Union[str, Any] = prepare_pegasus_inputs_dict(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) return config, inputs_dict def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : Optional[int] = TFPegasusModel(config=_lowerCamelCase).get_decoder() UpperCAmelCase__ : Optional[int] = inputs_dict["""input_ids"""] UpperCAmelCase__ : Any = input_ids[:1, :] UpperCAmelCase__ : List[str] = inputs_dict["""attention_mask"""][:1, :] UpperCAmelCase__ : int = inputs_dict["""head_mask"""] UpperCAmelCase__ : int = 1 # first forward pass UpperCAmelCase__ : int = model(_lowerCamelCase , attention_mask=_lowerCamelCase , head_mask=_lowerCamelCase , use_cache=_lowerCamelCase) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase__ : int = ids_tensor((self.batch_size, 3) , config.vocab_size) UpperCAmelCase__ : int = tf.cast(ids_tensor((self.batch_size, 3) , 2) , tf.inta) # append to next input_ids and UpperCAmelCase__ : Tuple = tf.concat([input_ids, next_tokens] , axis=-1) UpperCAmelCase__ : str = tf.concat([attention_mask, next_attn_mask] , axis=-1) UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase)[0] UpperCAmelCase__ : Any = model(_lowerCamelCase , attention_mask=_lowerCamelCase , past_key_values=_lowerCamelCase)[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1]) # select random slice UpperCAmelCase__ : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1])) UpperCAmelCase__ : int = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase__ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_lowerCamelCase , _lowerCamelCase , rtol=1e-3) def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , ): if attention_mask is None: UpperCAmelCase__ : Any = tf.cast(tf.math.not_equal(UpperCamelCase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase__ : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase__ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase__ : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _snake_case ( a__ , a__ , unittest.TestCase ): lowerCAmelCase :str = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () lowerCAmelCase :Any = (TFPegasusForConditionalGeneration,) if is_tf_available() else () lowerCAmelCase :List[str] = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) lowerCAmelCase :int = True lowerCAmelCase :Optional[int] = False lowerCAmelCase :Dict = False def snake_case__ ( self): UpperCAmelCase__ : List[str] = TFPegasusModelTester(self) UpperCAmelCase__ : Dict = ConfigTester(self , config_class=_lowerCamelCase) def snake_case__ ( self): self.config_tester.run_common_tests() def snake_case__ ( self): UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowerCamelCase) @require_sentencepiece @require_tokenizers @require_tf class _snake_case ( unittest.TestCase ): lowerCAmelCase :Dict = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] lowerCAmelCase :Any = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers lowerCAmelCase :Tuple = '''google/pegasus-xsum''' @cached_property def snake_case__ ( self): return AutoTokenizer.from_pretrained(self.model_name) @cached_property def snake_case__ ( self): UpperCAmelCase__ : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def snake_case__ ( self , **_lowerCamelCase): UpperCAmelCase__ : Dict = self.translate_src_text(**_lowerCamelCase) assert self.expected_text == generated_words def snake_case__ ( self , **_lowerCamelCase): UpperCAmelCase__ : List[Any] = self.tokenizer(self.src_text , **_lowerCamelCase , padding=_lowerCamelCase , return_tensors="""tf""") UpperCAmelCase__ : Dict = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=_lowerCamelCase , ) UpperCAmelCase__ : Optional[Any] = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_lowerCamelCase) return generated_words @slow def snake_case__ ( self): self._assert_generated_batch_equal_expected()
283
'''simple docstring''' from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _snake_case : lowerCAmelCase :str = PegasusConfig lowerCAmelCase :Optional[int] = {} lowerCAmelCase :Dict = '''gelu''' def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=False , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=2 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=40 , _lowerCamelCase=2 , _lowerCamelCase=1 , _lowerCamelCase=0 , ): UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : Union[str, Any] = seq_length UpperCAmelCase__ : str = is_training UpperCAmelCase__ : Dict = use_labels UpperCAmelCase__ : Optional[int] = vocab_size UpperCAmelCase__ : int = hidden_size UpperCAmelCase__ : Union[str, Any] = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : Optional[Any] = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Any = pad_token_id UpperCAmelCase__ : Dict = bos_token_id def snake_case__ ( self): UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) UpperCAmelCase__ : str = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) UpperCAmelCase__ : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1) UpperCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase__ : Any = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase__ : Union[str, Any] = prepare_pegasus_inputs_dict(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) return config, inputs_dict def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : Optional[int] = TFPegasusModel(config=_lowerCamelCase).get_decoder() UpperCAmelCase__ : Optional[int] = inputs_dict["""input_ids"""] UpperCAmelCase__ : Any = input_ids[:1, :] UpperCAmelCase__ : List[str] = inputs_dict["""attention_mask"""][:1, :] UpperCAmelCase__ : int = inputs_dict["""head_mask"""] UpperCAmelCase__ : int = 1 # first forward pass UpperCAmelCase__ : int = model(_lowerCamelCase , attention_mask=_lowerCamelCase , head_mask=_lowerCamelCase , use_cache=_lowerCamelCase) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase__ : int = ids_tensor((self.batch_size, 3) , config.vocab_size) UpperCAmelCase__ : int = tf.cast(ids_tensor((self.batch_size, 3) , 2) , tf.inta) # append to next input_ids and UpperCAmelCase__ : Tuple = tf.concat([input_ids, next_tokens] , axis=-1) UpperCAmelCase__ : str = tf.concat([attention_mask, next_attn_mask] , axis=-1) UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase)[0] UpperCAmelCase__ : Any = model(_lowerCamelCase , attention_mask=_lowerCamelCase , past_key_values=_lowerCamelCase)[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1]) # select random slice UpperCAmelCase__ : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1])) UpperCAmelCase__ : int = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase__ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_lowerCamelCase , _lowerCamelCase , rtol=1e-3) def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , ): if attention_mask is None: UpperCAmelCase__ : Any = tf.cast(tf.math.not_equal(UpperCamelCase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase__ : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase__ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase__ : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _snake_case ( a__ , a__ , unittest.TestCase ): lowerCAmelCase :str = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () lowerCAmelCase :Any = (TFPegasusForConditionalGeneration,) if is_tf_available() else () lowerCAmelCase :List[str] = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) lowerCAmelCase :int = True lowerCAmelCase :Optional[int] = False lowerCAmelCase :Dict = False def snake_case__ ( self): UpperCAmelCase__ : List[str] = TFPegasusModelTester(self) UpperCAmelCase__ : Dict = ConfigTester(self , config_class=_lowerCamelCase) def snake_case__ ( self): self.config_tester.run_common_tests() def snake_case__ ( self): UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowerCamelCase) @require_sentencepiece @require_tokenizers @require_tf class _snake_case ( unittest.TestCase ): lowerCAmelCase :Dict = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] lowerCAmelCase :Any = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers lowerCAmelCase :Tuple = '''google/pegasus-xsum''' @cached_property def snake_case__ ( self): return AutoTokenizer.from_pretrained(self.model_name) @cached_property def snake_case__ ( self): UpperCAmelCase__ : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def snake_case__ ( self , **_lowerCamelCase): UpperCAmelCase__ : Dict = self.translate_src_text(**_lowerCamelCase) assert self.expected_text == generated_words def snake_case__ ( self , **_lowerCamelCase): UpperCAmelCase__ : List[Any] = self.tokenizer(self.src_text , **_lowerCamelCase , padding=_lowerCamelCase , return_tensors="""tf""") UpperCAmelCase__ : Dict = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=_lowerCamelCase , ) UpperCAmelCase__ : Optional[Any] = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_lowerCamelCase) return generated_words @slow def snake_case__ ( self): self._assert_generated_batch_equal_expected()
283
1
import argparse import logging import pickle from collections import Counter logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) SCREAMING_SNAKE_CASE_ = logging.getLogger(__name__) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser( description="""Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)""" ) parser.add_argument( """--data_file""", type=str, default="""data/dump.bert-base-uncased.pickle""", help="""The binarized dataset.""" ) parser.add_argument( """--token_counts_dump""", type=str, default="""data/token_counts.bert-base-uncased.pickle""", help="""The dump file.""" ) parser.add_argument("""--vocab_size""", default=3_0_5_2_2, type=int) SCREAMING_SNAKE_CASE_ = parser.parse_args() logger.info(F'''Loading data from {args.data_file}''') with open(args.data_file, """rb""") as fp: SCREAMING_SNAKE_CASE_ = pickle.load(fp) logger.info("""Counting occurrences for MLM.""") SCREAMING_SNAKE_CASE_ = Counter() for tk_ids in data: counter.update(tk_ids) SCREAMING_SNAKE_CASE_ = [0] * args.vocab_size for k, v in counter.items(): SCREAMING_SNAKE_CASE_ = v logger.info(F'''Dump to {args.token_counts_dump}''') with open(args.token_counts_dump, """wb""") as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
296
"""simple docstring""" from collections import defaultdict from math import ceil, sqrt def __SCREAMING_SNAKE_CASE ( A_ = 1_00_00_00 , A_ = 10 ): lowerCAmelCase__ : defaultdict = defaultdict(A_ ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: lowerCAmelCase__ : int = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: lowerCAmelCase__ : Tuple = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(A_ , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(F'''{solution() = }''')
106
0
"""simple docstring""" import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __A : """simple docstring""" def __init__( self , __A , __A=13 , __A=32 , __A=2 , __A=3 , __A=16 , __A=[32, 64, 128] , __A=[1, 2, 1] , __A=[2, 2, 4] , __A=2 , __A=2.0 , __A=True , __A=0.0 , __A=0.0 , __A=0.1 , __A="gelu" , __A=False , __A=True , __A=0.02 , __A=1E-5 , __A=True , __A=None , __A=True , __A=10 , __A=8 , __A=["stage1", "stage2"] , __A=[1, 2] , ) -> int: a =parent a =batch_size a =image_size a =patch_size a =num_channels a =embed_dim a =hidden_sizes a =depths a =num_heads a =window_size a =mlp_ratio a =qkv_bias a =hidden_dropout_prob a =attention_probs_dropout_prob a =drop_path_rate a =hidden_act a =use_absolute_embeddings a =patch_norm a =layer_norm_eps a =initializer_range a =is_training a =scope a =use_labels a =type_sequence_label_size a =encoder_stride a =out_features a =out_indices def SCREAMING_SNAKE_CASE ( self ) -> int: a =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a =None if self.use_labels: a =ids_tensor([self.batch_size] , self.type_sequence_label_size ) a =self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A ) -> int: a =FocalNetModel(config=__A ) model.to(__A ) model.eval() a =model(__A ) a =((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) a =int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A ) -> int: a =FocalNetBackbone(config=__A ) model.to(__A ) model.eval() a =model(__A ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None a =None a =FocalNetBackbone(config=__A ) model.to(__A ) model.eval() a =model(__A ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A ) -> Any: a =FocalNetForMaskedImageModeling(config=__A ) model.to(__A ) model.eval() a =model(__A ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images a =1 a =FocalNetForMaskedImageModeling(__A ) model.to(__A ) model.eval() a =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) a =model(__A ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A ) -> str: a =self.type_sequence_label_size a =FocalNetForImageClassification(__A ) model.to(__A ) model.eval() a =model(__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images a =1 a =FocalNetForImageClassification(__A ) model.to(__A ) model.eval() a =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) a =model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =self.prepare_config_and_inputs() a , a , a =config_and_inputs a ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __A ( _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE, unittest.TestCase ): """simple docstring""" __lowerCAmelCase = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) __lowerCAmelCase = ( {"feature-extraction": FocalNetModel, "image-classification": FocalNetForImageClassification} if is_torch_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def SCREAMING_SNAKE_CASE ( self ) -> str: a =FocalNetModelTester(self ) a =ConfigTester(self , config_class=__A , embed_dim=37 , has_text_modality=__A ) def SCREAMING_SNAKE_CASE ( self ) -> int: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self ) -> int: return def SCREAMING_SNAKE_CASE ( self ) -> str: a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__A ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__A ) def SCREAMING_SNAKE_CASE ( self ) -> str: a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__A ) @unittest.skip(reason='''FocalNet does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: pass @unittest.skip(reason='''FocalNet does not use feedforward chunking''' ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: pass def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a , a =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: a =model_class(__A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) a =model.get_output_embeddings() self.assertTrue(x is None or isinstance(__A , nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a , a =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: a =model_class(__A ) a =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a =[*signature.parameters.keys()] a =['''pixel_values'''] self.assertListEqual(arg_names[:1] , __A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A ) -> List[Any]: a =model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): a =model(**self._prepare_for_class(__A , __A ) ) a =outputs.hidden_states a =getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__A ) , __A ) # FocalNet has a different seq_length a =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) a =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) a =outputs.reshaped_hidden_states self.assertEqual(len(__A ) , __A ) a , a , a , a =reshaped_hidden_states[0].shape a =( reshaped_hidden_states[0].view(__A , __A , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def SCREAMING_SNAKE_CASE ( self ) -> int: a , a =self.model_tester.prepare_config_and_inputs_for_common() a =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: a =True self.check_hidden_states_output(__A , __A , __A , __A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a =True self.check_hidden_states_output(__A , __A , __A , __A ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a , a =self.model_tester.prepare_config_and_inputs_for_common() a =3 a =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) a =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) a =image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) a =image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: a =True self.check_hidden_states_output(__A , __A , __A , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a =True self.check_hidden_states_output(__A , __A , __A , (padded_height, padded_width) ) @slow def SCREAMING_SNAKE_CASE ( self ) -> Dict: for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a =FocalNetModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def SCREAMING_SNAKE_CASE ( self ) -> List[str]: a , a =self.model_tester.prepare_config_and_inputs_for_common() a =_config_zero_init(__A ) for model_class in self.all_model_classes: a =model_class(config=__A ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class __A ( unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self ) -> str: # TODO update organization return AutoImageProcessor.from_pretrained('''microsoft/focalnet-tiny''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE ( self ) -> Any: a =FocalNetForImageClassification.from_pretrained('''microsoft/focalnet-tiny''' ).to(__A ) a =self.default_image_processor a =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) a =image_processor(images=__A , return_tensors='''pt''' ).to(__A ) # forward pass with torch.no_grad(): a =model(**__A ) # verify the logits a =torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __A ) a =torch.tensor([0.2_166, -0.4_368, 0.2_191] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __A , atol=1E-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class __A ( _SCREAMING_SNAKE_CASE, unittest.TestCase ): """simple docstring""" __lowerCAmelCase = (FocalNetBackbone,) if is_torch_available() else () __lowerCAmelCase = FocalNetConfig __lowerCAmelCase = False def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: a =FocalNetModelTester(self )
351
"""simple docstring""" from __future__ import annotations def _A ( lowercase , lowercase , lowercase , ): """simple docstring""" if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
215
0
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller _A : Union[str, Any] = 3 def __magic_name__ ( __snake_case : int ) -> Any: print("Generating primitive root of p" ) while True: lowercase : int = random.randrange(3 , __SCREAMING_SNAKE_CASE ) if pow(__SCREAMING_SNAKE_CASE , 2 , __SCREAMING_SNAKE_CASE ) == 1: continue if pow(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) == 1: continue return g def __magic_name__ ( __snake_case : int ) -> Tuple: print("Generating prime p..." ) lowercase : Optional[int] = rabin_miller.generate_large_prime(__SCREAMING_SNAKE_CASE ) # select large prime number. lowercase : Optional[Any] = primitive_root(__SCREAMING_SNAKE_CASE ) # one primitive root on modulo p. lowercase : List[str] = random.randrange(3 , __SCREAMING_SNAKE_CASE ) # private_key -> have to be greater than 2 for safety. lowercase : Union[str, Any] = cryptomath.find_mod_inverse(pow(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) lowercase : Any = (key_size, e_a, e_a, p) lowercase : Dict = (key_size, d) return public_key, private_key def __magic_name__ ( __snake_case : str , __snake_case : int ) -> Tuple: if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print("\nWARNING:" ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" "Use a different name or delete these files and re-run this program." ) sys.exit() lowercase : List[str] = generate_key(__SCREAMING_SNAKE_CASE ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , "w" ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , "w" ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def __magic_name__ ( ) -> Optional[Any]: print("Making key files..." ) make_key_files("elgamal" , 2048 ) print("Key files generation successful" ) if __name__ == "__main__": main()
202
"""simple docstring""" import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) class UpperCamelCase ( enum.Enum ): lowercase = 0 lowercase = 1 @add_end_docstrings(lowercase_ ) class UpperCamelCase ( lowercase_ ): lowercase = 'generated' def __init__( self ,*__UpperCamelCase ,**__UpperCamelCase ) -> Tuple: '''simple docstring''' super().__init__(*__UpperCamelCase ,**__UpperCamelCase ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == 'tf' else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def _UpperCAmelCase ( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,**__UpperCamelCase ,) -> Optional[Any]: '''simple docstring''' lowercase_ : List[Any] = {} if truncation is not None: lowercase_ : int = truncation lowercase_ : Dict = generate_kwargs lowercase_ : List[Any] = {} if return_tensors is not None and return_type is None: lowercase_ : Union[str, Any] = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: lowercase_ : str = return_type if clean_up_tokenization_spaces is not None: lowercase_ : Dict = clean_up_tokenization_spaces if stop_sequence is not None: lowercase_ : Union[str, Any] = self.tokenizer.encode(__UpperCamelCase ,add_special_tokens=__UpperCamelCase ) if len(__UpperCamelCase ) > 1: warnings.warn( 'Stopping on a multiple token sequence is not yet supported on transformers. The first token of' ' the stop sequence will be used as the stop sequence string in the interim.' ) lowercase_ : Optional[int] = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' return True def _UpperCAmelCase ( self ,*__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : Dict = self.model.config.prefix if self.model.config.prefix is not None else '' if isinstance(args[0] ,__UpperCamelCase ): if self.tokenizer.pad_token_id is None: raise ValueError('Please make sure that the tokenizer has a pad_token_id when using a batch input' ) lowercase_ : str = ([prefix + arg for arg in args[0]],) lowercase_ : Union[str, Any] = True elif isinstance(args[0] ,__UpperCamelCase ): lowercase_ : Union[str, Any] = (prefix + args[0],) lowercase_ : Union[str, Any] = False else: raise ValueError( f''' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`''' ) lowercase_ : List[Any] = self.tokenizer(*__UpperCamelCase ,padding=__UpperCamelCase ,truncation=__UpperCamelCase ,return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self ,*__UpperCamelCase ,**__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : Optional[int] = super().__call__(*__UpperCamelCase ,**__UpperCamelCase ) if ( isinstance(args[0] ,__UpperCamelCase ) and all(isinstance(__UpperCamelCase ,__UpperCamelCase ) for el in args[0] ) and all(len(__UpperCamelCase ) == 1 for res in result ) ): return [res[0] for res in result] return result def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase=TruncationStrategy.DO_NOT_TRUNCATE ,**__UpperCamelCase ) -> List[str]: '''simple docstring''' lowercase_ : Any = self._parse_and_tokenize(__UpperCamelCase ,truncation=__UpperCamelCase ,**__UpperCamelCase ) return inputs def _UpperCAmelCase ( self ,__UpperCamelCase ,**__UpperCamelCase ) -> Optional[int]: '''simple docstring''' if self.framework == "pt": lowercase_ , lowercase_ : Optional[int] = model_inputs['input_ids'].shape elif self.framework == "tf": lowercase_ , lowercase_ : Union[str, Any] = tf.shape(model_inputs['input_ids'] ).numpy() lowercase_ : str = generate_kwargs.get('min_length' ,self.model.config.min_length ) lowercase_ : List[Any] = generate_kwargs.get('max_length' ,self.model.config.max_length ) self.check_inputs(__UpperCamelCase ,generate_kwargs['min_length'] ,generate_kwargs['max_length'] ) lowercase_ : Tuple = self.model.generate(**__UpperCamelCase ,**__UpperCamelCase ) lowercase_ : str = output_ids.shape[0] if self.framework == "pt": lowercase_ : List[Any] = output_ids.reshape(__UpperCamelCase ,out_b // in_b ,*output_ids.shape[1:] ) elif self.framework == "tf": lowercase_ : List[Any] = tf.reshape(__UpperCamelCase ,(in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase=ReturnType.TEXT ,__UpperCamelCase=False ) -> Dict: '''simple docstring''' lowercase_ : Dict = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: lowercase_ : List[Any] = {f'''{self.return_name}_token_ids''': output_ids} elif return_type == ReturnType.TEXT: lowercase_ : str = { f'''{self.return_name}_text''': self.tokenizer.decode( __UpperCamelCase ,skip_special_tokens=__UpperCamelCase ,clean_up_tokenization_spaces=__UpperCamelCase ,) } records.append(__UpperCamelCase ) return records @add_end_docstrings(lowercase_ ) class UpperCamelCase ( lowercase_ ): lowercase = 'summary' def __call__( self ,*__UpperCamelCase ,**__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' return super().__call__(*__UpperCamelCase ,**__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> bool: '''simple docstring''' if max_length < min_length: logger.warning(f'''Your min_length={min_length} must be inferior than your max_length={max_length}.''' ) if input_length < max_length: logger.warning( f'''Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ''' 'a summarization task, where outputs shorter than the input are typically wanted, you might ' f'''consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})''' ) @add_end_docstrings(lowercase_ ) class UpperCamelCase ( lowercase_ ): lowercase = 'translation' def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' if input_length > 0.9 * max_length: logger.warning( f'''Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ''' 'increasing your max_length manually, e.g. translator(\'...\', max_length=400)' ) return True def _UpperCAmelCase ( self ,*__UpperCamelCase ,__UpperCamelCase=TruncationStrategy.DO_NOT_TRUNCATE ,__UpperCamelCase=None ,__UpperCamelCase=None ) -> int: '''simple docstring''' if getattr(self.tokenizer ,'_build_translation_inputs' ,__UpperCamelCase ): return self.tokenizer._build_translation_inputs( *__UpperCamelCase ,return_tensors=self.framework ,truncation=__UpperCamelCase ,src_lang=__UpperCamelCase ,tgt_lang=__UpperCamelCase ) else: return super()._parse_and_tokenize(*__UpperCamelCase ,truncation=__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,**__UpperCamelCase ) -> str: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ : int = super()._sanitize_parameters(**__UpperCamelCase ) if src_lang is not None: lowercase_ : str = src_lang if tgt_lang is not None: lowercase_ : Optional[Any] = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. lowercase_ : Tuple = kwargs.get('task' ,self.task ) lowercase_ : List[str] = task.split('_' ) if task and len(__UpperCamelCase ) == 4: # translation, XX, to YY lowercase_ : Union[str, Any] = items[1] lowercase_ : Tuple = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self ,*__UpperCamelCase ,**__UpperCamelCase ) -> Dict: '''simple docstring''' return super().__call__(*__UpperCamelCase ,**__UpperCamelCase )
213
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase_ ) class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ : str = field(default='''image-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) UpperCamelCase__ : ClassVar[Features] = Features({'''image''': Image()} ) UpperCamelCase__ : ClassVar[Features] = Features({'''labels''': ClassLabel} ) UpperCamelCase__ : str = "image" UpperCamelCase__ : str = "labels" def _A ( self , _A ): '''simple docstring''' if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , _A ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) __SCREAMING_SNAKE_CASE = copy.deepcopy(self ) __SCREAMING_SNAKE_CASE = self.label_schema.copy() __SCREAMING_SNAKE_CASE = features[self.label_column] __SCREAMING_SNAKE_CASE = label_schema return task_template @property def _A ( self ): '''simple docstring''' return { self.image_column: "image", self.label_column: "labels", }
118
from collections import deque from .hash_table import HashTable class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' def __init__( self , *_A , **_A ): '''simple docstring''' super().__init__(*_A , **_A ) def _A ( self , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_A ) __SCREAMING_SNAKE_CASE = self.values[key] def _A ( self ): '''simple docstring''' return ( sum(self.charge_factor - len(_A ) for slot in self.values ) / self.size_table * self.charge_factor ) def _A ( self , _A , _A=None ): '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_A ) == 0 ): return key return super()._collision_resolution(_A , _A )
118
1
def __magic_name__ ( __a : list[list[int]] , __a : int , __a : int , __a : set ): '''simple docstring''' UpperCamelCase__ = len(__a ), len(grid[0] ) if ( min(__a , __a ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) UpperCamelCase__ = 0 count += depth_first_search(__a , row + 1 , __a , __a ) count += depth_first_search(__a , row - 1 , __a , __a ) count += depth_first_search(__a , __a , col + 1 , __a ) count += depth_first_search(__a , __a , col - 1 , __a ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
244
"""simple docstring""" import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging __snake_case = logging.get_logger(__name__) def __lowerCAmelCase ( ) -> str: """simple docstring""" snake_case : Dict = os.getenv("SM_HP_MP_PARAMETERS" , "{}" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. snake_case : Optional[int] = json.loads(lowercase ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. snake_case : Optional[int] = os.getenv("SM_FRAMEWORK_PARAMS" , "{}" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". snake_case : Any = json.loads(lowercase ) if not mpi_options.get("sagemaker_mpi_enabled" , lowercase ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("smdistributed" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class _lowerCAmelCase ( snake_case_ ): __UpperCAmelCase : str = field( default='''''' , metadata={'''help''': '''Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'''} , ) def lowerCamelCase ( self ) -> List[Any]: '''simple docstring''' super().__post_init__() warnings.warn( "`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use " "`TrainingArguments` instead." , UpperCamelCase__ , ) @cached_property def lowerCamelCase ( self ) -> "torch.device": '''simple docstring''' logger.info("PyTorch: setting up devices" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( "torch.distributed process group is initialized, but local_rank == -1. " "In order to use Torch DDP, launch your script with `python -m torch.distributed.launch" ) if self.no_cuda: snake_case : Optional[Any] = torch.device("cpu" ) snake_case : List[Any] = 0 elif is_sagemaker_model_parallel_available(): snake_case : Tuple = smp.local_rank() snake_case : int = torch.device("cuda" , UpperCamelCase__ ) snake_case : Dict = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="smddp" , timeout=self.ddp_timeout_delta ) snake_case : Any = int(os.getenv("SMDATAPARALLEL_LOCAL_RANK" ) ) snake_case : Optional[Any] = torch.device("cuda" , self.local_rank ) snake_case : str = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 snake_case : List[str] = torch.device("cuda:0" if torch.cuda.is_available() else "cpu" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. snake_case : Optional[Any] = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="nccl" , timeout=self.ddp_timeout_delta ) snake_case : Any = torch.device("cuda" , self.local_rank ) snake_case : Dict = 1 if device.type == "cuda": torch.cuda.set_device(UpperCamelCase__ ) return device @property def lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def lowerCamelCase ( self ) -> List[str]: '''simple docstring''' return not is_sagemaker_model_parallel_available() @property def lowerCamelCase ( self ) -> str: '''simple docstring''' return False
203
0
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCAmelCase__ : Optional[Any] =datasets.logging.get_logger(__name__) UpperCAmelCase__ : List[str] ='''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' UpperCAmelCase__ : str ='''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' UpperCAmelCase__ : str =''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=False , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase="dummy_doc" ) -> List[str]: lowerCamelCase ={doc: key_lines} lowerCamelCase ={doc: sys_lines} lowerCamelCase ={} lowerCamelCase =0 lowerCamelCase =0 lowerCamelCase =0 lowerCamelCase =0 lowerCamelCase =0 lowerCamelCase =0 lowerCamelCase , lowerCamelCase =reader.get_doc_mentions(_UpperCAmelCase , key_doc_lines[doc] , _UpperCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: lowerCamelCase =reader.set_annotated_parse_trees(_UpperCAmelCase , key_doc_lines[doc] , _UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase , lowerCamelCase =reader.get_doc_mentions(_UpperCAmelCase , sys_doc_lines[doc] , _UpperCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: lowerCamelCase =reader.set_annotated_parse_trees(_UpperCAmelCase , key_doc_lines[doc] , _UpperCAmelCase , _UpperCAmelCase ) if remove_nested: lowerCamelCase , lowerCamelCase =reader.remove_nested_coref_mentions(_UpperCAmelCase , _UpperCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowerCamelCase , lowerCamelCase =reader.remove_nested_coref_mentions(_UpperCAmelCase , _UpperCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowerCamelCase =reader.get_mention_assignments(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase =reader.get_mention_assignments(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( """Number of resulting singleton clusters in the key """ F"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( F"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ """files, respectively""" ) return doc_coref_infos def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase =get_coref_infos(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase ={} lowerCamelCase =0 lowerCamelCase =0 for name, metric in metrics: lowerCamelCase , lowerCamelCase , lowerCamelCase =evaluator.evaluate_documents(_UpperCAmelCase , _UpperCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"""{name}/recall""": recall, F"""{name}/precision""": precision, F"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , F"""Recall: {recall * 1_00:.2f}""" , F""" Precision: {precision * 1_00:.2f}""" , F""" F1: {fa * 1_00:.2f}""" , ) if conll_subparts_num == 3: lowerCamelCase =(conll / 3) * 1_00 logger.info(F"""CoNLL score: {conll:.2f}""" ) output_scores.update({"""conll_score""": conll} ) return output_scores def _lowercase ( _UpperCAmelCase ) -> Any: lowerCamelCase =False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: lowerCamelCase =line.split()[5] if not parse_col == "-": lowerCamelCase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): def _snake_case ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=False , UpperCAmelCase_=False ): lowerCamelCase =[ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: lowerCamelCase =util.check_gold_parse_annotation(UpperCAmelCase_ ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowerCamelCase =evaluate( key_lines=UpperCAmelCase_ , sys_lines=UpperCAmelCase_ , metrics=UpperCAmelCase_ , NP_only=UpperCAmelCase_ , remove_nested=UpperCAmelCase_ , keep_singletons=UpperCAmelCase_ , min_span=UpperCAmelCase_ , ) return score
262
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def _lowercase ( _UpperCAmelCase = "isbn/0140328726" ) -> dict: lowerCamelCase =olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: lowerCamelCase =F"""{olid} is not a valid Open Library olid""" raise ValueError(_UpperCAmelCase ) return requests.get(F"""https://openlibrary.org/{new_olid}.json""" ).json() def _lowercase ( _UpperCAmelCase ) -> dict: lowerCamelCase ={ """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } lowerCamelCase ={better_key: ol_book_data[key] for key, better_key in desired_keys.items()} lowerCamelCase =[ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] lowerCamelCase =data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase =""", """.join(_UpperCAmelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: UpperCAmelCase__ : List[str] =input('''\nEnter the ISBN code to search (or \'quit\' to stop): ''').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: UpperCAmelCase__ : Dict =summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('''\n'''.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
262
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { "configuration_nezha": ["NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP", "NezhaConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[str] = [ "NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST", "NezhaForNextSentencePrediction", "NezhaForMaskedLM", "NezhaForPreTraining", "NezhaForMultipleChoice", "NezhaForQuestionAnswering", "NezhaForSequenceClassification", "NezhaForTokenClassification", "NezhaModel", "NezhaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys UpperCAmelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
136
"""simple docstring""" import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors UpperCAmelCase : Any = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "sequence-classification" def __init__( self : Optional[Any] , lowerCAmelCase_ : int): """simple docstring""" if type(lowerCAmelCase_) == dict: lowercase_ = Namespace(**lowerCAmelCase_) lowercase_ = glue_output_modes[hparams.task] lowercase_ = glue_tasks_num_labels[hparams.task] super().__init__(lowerCAmelCase_ , lowerCAmelCase_ , self.mode) def _UpperCAmelCase ( self : Optional[int] , **lowerCAmelCase_ : Optional[int]): """simple docstring""" return self.model(**lowerCAmelCase_) def _UpperCAmelCase ( self : List[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[Any]): """simple docstring""" lowercase_ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowercase_ = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowercase_ = self(**lowerCAmelCase_) lowercase_ = outputs[0] lowercase_ = self.trainer.lr_schedulers[0]["""scheduler"""] lowercase_ = {"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def _UpperCAmelCase ( self : List[str]): """simple docstring""" lowercase_ = self.hparams lowercase_ = processors[args.task]() lowercase_ = processor.get_labels() for mode in ["train", "dev"]: lowercase_ = self._feature_file(lowerCAmelCase_) if os.path.exists(lowerCAmelCase_) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , lowerCAmelCase_) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir) lowercase_ = ( processor.get_dev_examples(args.data_dir) if mode == """dev""" else processor.get_train_examples(args.data_dir) ) lowercase_ = convert_examples_to_features( lowerCAmelCase_ , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , lowerCAmelCase_) torch.save(lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : int , lowerCAmelCase_ : bool = False): """simple docstring""" lowercase_ = """dev""" if mode == """test""" else mode lowercase_ = self._feature_file(lowerCAmelCase_) logger.info("""Loading features from cached file %s""" , lowerCAmelCase_) lowercase_ = torch.load(lowerCAmelCase_) lowercase_ = torch.tensor([f.input_ids for f in features] , dtype=torch.long) lowercase_ = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) lowercase_ = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) if self.hparams.glue_output_mode == "classification": lowercase_ = torch.tensor([f.label for f in features] , dtype=torch.long) elif self.hparams.glue_output_mode == "regression": lowercase_ = torch.tensor([f.label for f in features] , dtype=torch.float) return DataLoader( TensorDataset(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) , batch_size=lowerCAmelCase_ , shuffle=lowerCAmelCase_ , ) def _UpperCAmelCase ( self : Tuple , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[Any]): """simple docstring""" lowercase_ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowercase_ = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowercase_ = self(**lowerCAmelCase_) lowercase_ , lowercase_ = outputs[:2] lowercase_ = logits.detach().cpu().numpy() lowercase_ = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _UpperCAmelCase ( self : str , lowerCAmelCase_ : int): """simple docstring""" lowercase_ = torch.stack([x["""val_loss"""] for x in outputs]).mean().detach().cpu().item() lowercase_ = np.concatenate([x["""pred"""] for x in outputs] , axis=0) if self.hparams.glue_output_mode == "classification": lowercase_ = np.argmax(lowerCAmelCase_ , axis=1) elif self.hparams.glue_output_mode == "regression": lowercase_ = np.squeeze(lowerCAmelCase_) lowercase_ = np.concatenate([x["""target"""] for x in outputs] , axis=0) lowercase_ = [[] for _ in range(out_label_ids.shape[0])] lowercase_ = [[] for _ in range(out_label_ids.shape[0])] lowercase_ = {**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , lowerCAmelCase_ , lowerCAmelCase_)} lowercase_ = dict(results.items()) lowercase_ = results return ret, preds_list, out_label_list def _UpperCAmelCase ( self : int , lowerCAmelCase_ : list): """simple docstring""" lowercase_ , lowercase_ , lowercase_ = self._eval_end(lowerCAmelCase_) lowercase_ = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _UpperCAmelCase ( self : Tuple , lowerCAmelCase_ : int): """simple docstring""" lowercase_ , lowercase_ , lowercase_ = self._eval_end(lowerCAmelCase_) lowercase_ = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _UpperCAmelCase ( lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : str): """simple docstring""" BaseTransformer.add_model_specific_args(lowerCAmelCase_ , lowerCAmelCase_) parser.add_argument( """--max_seq_length""" , default=1_2_8 , type=lowerCAmelCase_ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=lowerCAmelCase_ , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""") return parser def _SCREAMING_SNAKE_CASE () -> str: '''simple docstring''' lowercase_ = argparse.ArgumentParser() add_generic_args(__lowerCAmelCase , os.getcwd() ) lowercase_ = GLUETransformer.add_model_specific_args(__lowerCAmelCase , os.getcwd() ) lowercase_ = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: lowercase_ = os.path.join( """./results""" , F'''{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}''' , ) os.makedirs(args.output_dir ) lowercase_ = GLUETransformer(__lowerCAmelCase ) lowercase_ = generic_train(__lowerCAmelCase , __lowerCAmelCase ) # Optionally, predict on dev set and write to output_dir if args.do_predict: lowercase_ = sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=__lowerCAmelCase ) ) lowercase_ = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(__lowerCAmelCase ) if __name__ == "__main__": main()
136
1
from manim import * class lowercase ( _UpperCAmelCase ): def __snake_case( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = Rectangle(height=0.5 , width=0.5 ) SCREAMING_SNAKE_CASE = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] SCREAMING_SNAKE_CASE = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = Text("CPU" , font_size=24 ) SCREAMING_SNAKE_CASE = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) SCREAMING_SNAKE_CASE = [mem.copy() for i in range(4 )] SCREAMING_SNAKE_CASE = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = Text("GPU" , font_size=24 ) SCREAMING_SNAKE_CASE = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] SCREAMING_SNAKE_CASE = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = Text("Model" , font_size=24 ) SCREAMING_SNAKE_CASE = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) SCREAMING_SNAKE_CASE = Rectangle(height=0.4_6 / 4 , width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] SCREAMING_SNAKE_CASE = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) SCREAMING_SNAKE_CASE = Text("Loaded Checkpoint" , font_size=24 ) SCREAMING_SNAKE_CASE = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) SCREAMING_SNAKE_CASE = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) SCREAMING_SNAKE_CASE = MarkupText( F"<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) SCREAMING_SNAKE_CASE = MarkupText( F"<span fgcolor=\'{BLUE}\'>●</span> Checkpoint" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) SCREAMING_SNAKE_CASE = MarkupText( F"Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(lowercase_ ): SCREAMING_SNAKE_CASE = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) SCREAMING_SNAKE_CASE = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
371
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase : lowercase__ : Dict = LEDConfig lowercase__ : List[str] = {} lowercase__ : Union[str, Any] = """gelu""" def __init__( self : Tuple , _UpperCamelCase : List[str] , _UpperCamelCase : Dict=13 , _UpperCamelCase : Optional[int]=7 , _UpperCamelCase : int=True , _UpperCamelCase : List[Any]=False , _UpperCamelCase : Dict=99 , _UpperCamelCase : Optional[Any]=32 , _UpperCamelCase : Any=2 , _UpperCamelCase : List[str]=4 , _UpperCamelCase : Union[str, Any]=37 , _UpperCamelCase : str=0.1 , _UpperCamelCase : List[Any]=0.1 , _UpperCamelCase : Union[str, Any]=20 , _UpperCamelCase : str=2 , _UpperCamelCase : Optional[Any]=1 , _UpperCamelCase : Optional[int]=0 , _UpperCamelCase : int=4 , ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = seq_length SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = eos_token_id SCREAMING_SNAKE_CASE = pad_token_id SCREAMING_SNAKE_CASE = bos_token_id SCREAMING_SNAKE_CASE = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after SCREAMING_SNAKE_CASE = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests SCREAMING_SNAKE_CASE = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __snake_case( self : int ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) SCREAMING_SNAKE_CASE = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) SCREAMING_SNAKE_CASE = tf.concat([input_ids, eos_tensor] , axis=1 ) SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) SCREAMING_SNAKE_CASE = prepare_led_inputs_dict(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = tf.concat( [tf.zeros_like(_UpperCamelCase )[:, :-1], tf.ones_like(_UpperCamelCase )[:, -1:]] , axis=-1 , ) SCREAMING_SNAKE_CASE = global_attention_mask return config, inputs_dict def __snake_case( self : Any , _UpperCamelCase : Optional[int] , _UpperCamelCase : Tuple ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = TFLEDModel(config=_UpperCamelCase ).get_decoder() SCREAMING_SNAKE_CASE = inputs_dict["input_ids"] SCREAMING_SNAKE_CASE = input_ids[:1, :] SCREAMING_SNAKE_CASE = inputs_dict["attention_mask"][:1, :] SCREAMING_SNAKE_CASE = 1 # first forward pass SCREAMING_SNAKE_CASE = model(_UpperCamelCase , attention_mask=_UpperCamelCase , use_cache=_UpperCamelCase ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and SCREAMING_SNAKE_CASE = tf.concat([input_ids, next_tokens] , axis=-1 ) SCREAMING_SNAKE_CASE = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) SCREAMING_SNAKE_CASE = model(_UpperCamelCase , attention_mask=_UpperCamelCase )[0] SCREAMING_SNAKE_CASE = model(_UpperCamelCase , attention_mask=_UpperCamelCase , past_key_values=_UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice SCREAMING_SNAKE_CASE = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) SCREAMING_SNAKE_CASE = output_from_no_past[:, -3:, random_slice_idx] SCREAMING_SNAKE_CASE = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_UpperCamelCase , _UpperCamelCase , rtol=1e-3 ) def __lowerCamelCase (UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Union[str, Any]=None , ): if attention_mask is None: SCREAMING_SNAKE_CASE = tf.cast(tf.math.not_equal(UpperCAmelCase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: SCREAMING_SNAKE_CASE = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase ( a , a , unittest.TestCase ): lowercase__ : Optional[int] = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () lowercase__ : List[Any] = (TFLEDForConditionalGeneration,) if is_tf_available() else () lowercase__ : int = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ : List[Any] = True lowercase__ : List[str] = False lowercase__ : List[str] = False lowercase__ : Union[str, Any] = False def __snake_case( self : Tuple ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = TFLEDModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=_UpperCamelCase ) def __snake_case( self : List[Any] ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def __snake_case( self : List[str] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_UpperCamelCase ) def __snake_case( self : Dict ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE = tf.zeros_like(inputs_dict["attention_mask"] ) SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = self.model_tester.seq_length SCREAMING_SNAKE_CASE = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_UpperCamelCase : Dict ): SCREAMING_SNAKE_CASE = outputs.decoder_attentions self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_UpperCamelCase : Optional[Any] ): SCREAMING_SNAKE_CASE = [t.numpy() for t in outputs.encoder_attentions] SCREAMING_SNAKE_CASE = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) SCREAMING_SNAKE_CASE = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) SCREAMING_SNAKE_CASE = len(_UpperCamelCase ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) if self.is_encoder_decoder: SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) SCREAMING_SNAKE_CASE = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_decoder_attentions_output(_UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) SCREAMING_SNAKE_CASE = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) # Check attention is always last and order is fine SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) SCREAMING_SNAKE_CASE = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def __snake_case( self : Optional[Any] ) -> Tuple: '''simple docstring''' pass def __snake_case( self : str ) -> str: '''simple docstring''' pass def __lowerCamelCase (UpperCAmelCase__ : Optional[int] ): return tf.constant(UpperCAmelCase__ , dtype=tf.intaa ) _lowerCamelCase : str = 1e-4 @slow @require_tf class lowercase ( unittest.TestCase ): def __snake_case( self : int ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here SCREAMING_SNAKE_CASE = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) SCREAMING_SNAKE_CASE = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) SCREAMING_SNAKE_CASE = prepare_led_inputs_dict(model.config , _UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = model(**_UpperCamelCase )[0] SCREAMING_SNAKE_CASE = (1, 1_024, 768) self.assertEqual(output.shape , _UpperCamelCase ) # change to expected output here SCREAMING_SNAKE_CASE = tf.convert_to_tensor( [[2.3_0_5_0, 2.8_2_7_9, 0.6_5_3_1], [-1.8_4_5_7, -0.1_4_5_5, -3.5_6_6_1], [-1.0_1_8_6, 0.4_5_8_6, -2.2_0_4_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCamelCase , atol=1e-3 ) def __snake_case( self : Any ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here SCREAMING_SNAKE_CASE = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) SCREAMING_SNAKE_CASE = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) SCREAMING_SNAKE_CASE = prepare_led_inputs_dict(model.config , _UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = model(**_UpperCamelCase )[0] SCREAMING_SNAKE_CASE = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , _UpperCamelCase ) # change to expected output here SCREAMING_SNAKE_CASE = tf.convert_to_tensor( [[3_3.6_5_0_7, 6.4_5_7_2, 1_6.8_0_8_9], [5.8_7_3_9, -2.4_2_3_8, 1_1.2_9_0_2], [-3.2_1_3_9, -4.3_1_4_9, 4.2_7_8_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCamelCase , atol=1e-3 , rtol=1e-3 )
206
0
'''simple docstring''' import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _UpperCamelCase ( lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' _A : List[str] = TransfoXLTokenizer _A : List[Any] = False _A : Dict = False def UpperCamelCase__ ( self : Optional[Any] ): """simple docstring""" super().setUp() __SCREAMING_SNAKE_CASE : Union[str, Any] = [ """<unk>""", """[CLS]""", """[SEP]""", """want""", """unwanted""", """wa""", """un""", """running""", """,""", """low""", """l""", ] __SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def UpperCamelCase__ ( self : Optional[Any] , **lowerCAmelCase__ : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def UpperCamelCase__ ( self : Tuple , lowerCAmelCase__ : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = """<unk> UNwanted , running""" __SCREAMING_SNAKE_CASE : List[Any] = """<unk> unwanted, running""" return input_text, output_text def UpperCamelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = tokenizer.tokenize("""<unk> UNwanted , running""" ) self.assertListEqual(lowerCAmelCase__ , ["""<unk>""", """unwanted""", """,""", """running"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [0, 4, 8, 7] ) def UpperCamelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = TransfoXLTokenizer(lower_case=lowerCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) def UpperCamelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = TransfoXLTokenizer(lower_case=lowerCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def UpperCamelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = TransfoXLTokenizer(lower_case=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = """Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?""" __SCREAMING_SNAKE_CASE : Optional[int] = [ """Hello""", """(""", """bracket""", """)""", """and""", """side""", """@-@""", """scrolled""", """[""", """and""", """]""", """Henry""", """'s""", """$""", """5""", """@,@""", """000""", """with""", """3""", """@.@""", """34""", """m""", """.""", """What""", """'s""", """up""", """!""", """?""", ] self.assertListEqual(tokenizer.tokenize(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = self.get_tokenizer() __SCREAMING_SNAKE_CASE : str = len(lowerCAmelCase__ ) tokenizer.add_tokens(["""new1""", """new2"""] ) tokenizer.move_added_token("""new1""" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(lowerCAmelCase__ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("""new1""" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , """new1""" )
112
'''simple docstring''' from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase__ : Dict = [ '''python''', '''tqdm''', '''regex''', '''requests''', '''packaging''', '''filelock''', '''numpy''', '''tokenizers''', '''huggingface-hub''', '''safetensors''', '''accelerate''', '''pyyaml''', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"can't find {pkg} in {deps.keys()}, check dependency_versions_table.py") def lowerCAmelCase_ ( _lowerCamelCase: List[Any] , _lowerCamelCase: Optional[int]=None ): require_version(deps[pkg] , _lowerCamelCase )
112
1
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 42 @flax_register_to_config class A ( nn.Module , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 32 lowerCamelCase = 4 lowerCamelCase = 4 lowerCamelCase = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowerCamelCase = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") lowerCamelCase = False lowerCamelCase = (3_20, 6_40, 12_80, 12_80) lowerCamelCase = 2 lowerCamelCase = 8 lowerCamelCase = None lowerCamelCase = 12_80 lowerCamelCase = 0.0 lowerCamelCase = False lowerCamelCase = jnp.floataa lowerCamelCase = True lowerCamelCase = 0 lowerCamelCase = False def snake_case__ ( self : str,lowercase_ : jax.random.KeyArray )-> FrozenDict: '''simple docstring''' A__ = (1, self.in_channels, self.sample_size, self.sample_size) A__ = jnp.zeros(lowercase_,dtype=jnp.floataa ) A__ = jnp.ones((1,),dtype=jnp.intaa ) A__ = jnp.zeros((1, 1, self.cross_attention_dim),dtype=jnp.floataa ) A__ , A__ = jax.random.split(lowercase_ ) A__ = {'params': params_rng, 'dropout': dropout_rng} return self.init(lowercase_,lowercase_,lowercase_,lowercase_ )["params"] def snake_case__ ( self : List[Any] )-> Optional[int]: '''simple docstring''' A__ = self.block_out_channels A__ = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( 'At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.' ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. A__ = self.num_attention_heads or self.attention_head_dim # input A__ = nn.Conv( block_out_channels[0],kernel_size=(3, 3),strides=(1, 1),padding=((1, 1), (1, 1)),dtype=self.dtype,) # time A__ = FlaxTimesteps( block_out_channels[0],flip_sin_to_cos=self.flip_sin_to_cos,freq_shift=self.config.freq_shift ) A__ = FlaxTimestepEmbedding(lowercase_,dtype=self.dtype ) A__ = self.only_cross_attention if isinstance(lowercase_,lowercase_ ): A__ = (only_cross_attention,) * len(self.down_block_types ) if isinstance(lowercase_,lowercase_ ): A__ = (num_attention_heads,) * len(self.down_block_types ) # down A__ = [] A__ = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): A__ = output_channel A__ = block_out_channels[i] A__ = i == len(lowercase_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": A__ = FlaxCrossAttnDownBlockaD( in_channels=lowercase_,out_channels=lowercase_,dropout=self.dropout,num_layers=self.layers_per_block,num_attention_heads=num_attention_heads[i],add_downsample=not is_final_block,use_linear_projection=self.use_linear_projection,only_cross_attention=only_cross_attention[i],use_memory_efficient_attention=self.use_memory_efficient_attention,dtype=self.dtype,) else: A__ = FlaxDownBlockaD( in_channels=lowercase_,out_channels=lowercase_,dropout=self.dropout,num_layers=self.layers_per_block,add_downsample=not is_final_block,dtype=self.dtype,) down_blocks.append(lowercase_ ) A__ = down_blocks # mid A__ = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1],dropout=self.dropout,num_attention_heads=num_attention_heads[-1],use_linear_projection=self.use_linear_projection,use_memory_efficient_attention=self.use_memory_efficient_attention,dtype=self.dtype,) # up A__ = [] A__ = list(reversed(lowercase_ ) ) A__ = list(reversed(lowercase_ ) ) A__ = list(reversed(lowercase_ ) ) A__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): A__ = output_channel A__ = reversed_block_out_channels[i] A__ = reversed_block_out_channels[min(i + 1,len(lowercase_ ) - 1 )] A__ = i == len(lowercase_ ) - 1 if up_block_type == "CrossAttnUpBlock2D": A__ = FlaxCrossAttnUpBlockaD( in_channels=lowercase_,out_channels=lowercase_,prev_output_channel=lowercase_,num_layers=self.layers_per_block + 1,num_attention_heads=reversed_num_attention_heads[i],add_upsample=not is_final_block,dropout=self.dropout,use_linear_projection=self.use_linear_projection,only_cross_attention=only_cross_attention[i],use_memory_efficient_attention=self.use_memory_efficient_attention,dtype=self.dtype,) else: A__ = FlaxUpBlockaD( in_channels=lowercase_,out_channels=lowercase_,prev_output_channel=lowercase_,num_layers=self.layers_per_block + 1,add_upsample=not is_final_block,dropout=self.dropout,dtype=self.dtype,) up_blocks.append(lowercase_ ) A__ = output_channel A__ = up_blocks # out A__ = nn.GroupNorm(num_groups=3_2,epsilon=1E-5 ) A__ = nn.Conv( self.out_channels,kernel_size=(3, 3),strides=(1, 1),padding=((1, 1), (1, 1)),dtype=self.dtype,) def __call__( self : Dict,lowercase_ : Any,lowercase_ : Optional[int],lowercase_ : Tuple,lowercase_ : Tuple=None,lowercase_ : Tuple=None,lowercase_ : bool = True,lowercase_ : bool = False,)-> Union[FlaxUNetaDConditionOutput, Tuple]: '''simple docstring''' if not isinstance(lowercase_,jnp.ndarray ): A__ = jnp.array([timesteps],dtype=jnp.intaa ) elif isinstance(lowercase_,jnp.ndarray ) and len(timesteps.shape ) == 0: A__ = timesteps.astype(dtype=jnp.floataa ) A__ = jnp.expand_dims(lowercase_,0 ) A__ = self.time_proj(lowercase_ ) A__ = self.time_embedding(lowercase_ ) # 2. pre-process A__ = jnp.transpose(lowercase_,(0, 2, 3, 1) ) A__ = self.conv_in(lowercase_ ) # 3. down A__ = (sample,) for down_block in self.down_blocks: if isinstance(lowercase_,lowercase_ ): A__ , A__ = down_block(lowercase_,lowercase_,lowercase_,deterministic=not train ) else: A__ , A__ = down_block(lowercase_,lowercase_,deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: A__ = () for down_block_res_sample, down_block_additional_residual in zip( lowercase_,lowercase_ ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) A__ = new_down_block_res_samples # 4. mid A__ = self.mid_block(lowercase_,lowercase_,lowercase_,deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: A__ = down_block_res_samples[-(self.layers_per_block + 1) :] A__ = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(lowercase_,lowercase_ ): A__ = up_block( lowercase_,temb=lowercase_,encoder_hidden_states=lowercase_,res_hidden_states_tuple=lowercase_,deterministic=not train,) else: A__ = up_block(lowercase_,temb=lowercase_,res_hidden_states_tuple=lowercase_,deterministic=not train ) # 6. post-process A__ = self.conv_norm_out(lowercase_ ) A__ = nn.silu(lowercase_ ) A__ = self.conv_out(lowercase_ ) A__ = jnp.transpose(lowercase_,(0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=lowercase_ )
282
import numpy as np from transformers import Pipeline def _snake_case( SCREAMING_SNAKE_CASE__ : int ) -> int: '''simple docstring''' A__ = np.max(SCREAMING_SNAKE_CASE__ , axis=-1 , keepdims=SCREAMING_SNAKE_CASE__ ) A__ = np.exp(outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=SCREAMING_SNAKE_CASE__ ) class A ( _UpperCAmelCase ): """simple docstring""" def snake_case__ ( self : Dict,**lowercase_ : Tuple )-> Tuple: '''simple docstring''' A__ = {} if "second_text" in kwargs: A__ = kwargs['second_text'] return preprocess_kwargs, {}, {} def snake_case__ ( self : List[Any],lowercase_ : int,lowercase_ : Optional[int]=None )-> List[str]: '''simple docstring''' return self.tokenizer(lowercase_,text_pair=lowercase_,return_tensors=self.framework ) def snake_case__ ( self : str,lowercase_ : Dict )-> List[str]: '''simple docstring''' return self.model(**lowercase_ ) def snake_case__ ( self : Dict,lowercase_ : Optional[int] )-> Dict: '''simple docstring''' A__ = model_outputs.logits[0].numpy() A__ = softmax(lowercase_ ) A__ = np.argmax(lowercase_ ) A__ = self.model.config.idalabel[best_class] A__ = probabilities[best_class].item() A__ = logits.tolist() return {"label": label, "score": score, "logits": logits}
282
1
class _lowerCamelCase ( a ): """simple docstring""" pass class _lowerCamelCase ( a ): """simple docstring""" pass class _lowerCamelCase : """simple docstring""" def __init__( self ) -> Union[str, Any]: '''simple docstring''' __snake_case : int = [ [], [], [], ] def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase ) -> None: '''simple docstring''' try: if len(self.queues[priority] ) >= 100: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(UpperCAmelCase ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def UpperCAmelCase ( self ) -> int: '''simple docstring''' for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self ) -> str: '''simple docstring''' return "\n".join(F"""Priority {i}: {q}""" for i, q in enumerate(self.queues ) ) class _lowerCamelCase : """simple docstring""" def __init__( self ) -> Union[str, Any]: '''simple docstring''' __snake_case : Union[str, Any] = [] def UpperCAmelCase ( self , UpperCAmelCase ) -> None: '''simple docstring''' if len(self.queue ) == 100: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(UpperCAmelCase ) def UpperCAmelCase ( self ) -> int: '''simple docstring''' if not self.queue: raise UnderFlowError("The queue is empty" ) else: __snake_case : Optional[int] = min(self.queue ) self.queue.remove(UpperCAmelCase ) return data def __str__( self ) -> str: '''simple docstring''' return str(self.queue ) def lowerCAmelCase__( ) -> List[Any]: __snake_case : int = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(lowercase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(lowercase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def lowerCAmelCase__( ) -> int: __snake_case : str = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(lowercase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(lowercase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
326
import argparse import datetime def lowerCAmelCase__( lowercase : str ) -> str: __snake_case : int = { "0": "Sunday", "1": "Monday", "2": "Tuesday", "3": "Wednesday", "4": "Thursday", "5": "Friday", "6": "Saturday", } __snake_case : int = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(lowercase ) < 11: raise ValueError("Must be 10 characters long" ) # Get month __snake_case : int = int(date_input[0] + date_input[1] ) # Validate if not 0 < m < 13: raise ValueError("Month must be between 1 - 12" ) __snake_case : str = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError("Date separator must be '-' or '/'" ) # Get day __snake_case : int = int(date_input[3] + date_input[4] ) # Validate if not 0 < d < 32: raise ValueError("Date must be between 1 - 31" ) # Get second separator __snake_case : str = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError("Date separator must be '-' or '/'" ) # Get year __snake_case : int = int(date_input[6] + date_input[7] + date_input[8] + date_input[9] ) # Arbitrary year range if not 45 < y < 8500: raise ValueError( "Year out of range. There has to be some sort of limit...right?" ) # Get datetime obj for validation __snake_case : str = datetime.date(int(lowercase ) , int(lowercase ) , int(lowercase ) ) # Start math if m <= 2: __snake_case : Optional[Any] = y - 1 __snake_case : Tuple = m + 12 # maths var __snake_case : int = int(str(lowercase )[:2] ) __snake_case : int = int(str(lowercase )[2:] ) __snake_case : int = int(2.6 * m - 5.3_9 ) __snake_case : int = int(c / 4 ) __snake_case : int = int(k / 4 ) __snake_case : int = int(d + k ) __snake_case : int = int(t + u + v + x ) __snake_case : int = int(z - (2 * c) ) __snake_case : int = round(w % 7 ) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError("The date was evaluated incorrectly. Contact developer." ) # Response __snake_case : str = f"""Your date {date_input}, is a {days[str(lowercase )]}!""" return response if __name__ == "__main__": import doctest doctest.testmod() _UpperCamelCase = argparse.ArgumentParser( description=( '''Find out what day of the week nearly any date is or was. Enter ''' '''date as a string in the mm-dd-yyyy or mm/dd/yyyy format''' ) ) parser.add_argument( '''date_input''', type=str, help='''Date as a string (mm-dd-yyyy or mm/dd/yyyy)''' ) _UpperCamelCase = parser.parse_args() zeller(args.date_input)
326
1
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class UpperCAmelCase__ : """simple docstring""" def __init__( self : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Tuple=None , __lowerCamelCase : str="resnet50" , __lowerCamelCase : Optional[Any]=3 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : List[Any]=3 , __lowerCamelCase : Dict=True , __lowerCamelCase : Dict=True , ) -> Tuple: SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else [4] SCREAMING_SNAKE_CASE__ = stage_names SCREAMING_SNAKE_CASE__ = out_features SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = is_training def lowercase_ ( self : Dict ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, pixel_values def lowercase_ ( self : str ) -> str: return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def lowercase_ ( self : Optional[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = TimmBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def lowercase_ ( self : Dict ) -> int: SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = config_and_inputs SCREAMING_SNAKE_CASE__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch @require_timm class UpperCAmelCase__ ( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" a = (TimmBackbone,) if is_torch_available() else () a = {"feature-extraction": TimmBackbone} if is_torch_available() else {} a = False a = False a = False a = False def lowercase_ ( self : str ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = TimmBackboneModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def lowercase_ ( self : Union[str, Any] ) -> List[Any]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase_ ( self : Any ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = '''resnet18''' SCREAMING_SNAKE_CASE__ = '''microsoft/resnet-18''' SCREAMING_SNAKE_CASE__ = AutoBackbone.from_pretrained(__lowerCamelCase , use_timm_backbone=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = AutoBackbone.from_pretrained(__lowerCamelCase ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) SCREAMING_SNAKE_CASE__ = AutoBackbone.from_pretrained(__lowerCamelCase , use_timm_backbone=__lowerCamelCase , out_indices=[1, 2, 3] ) SCREAMING_SNAKE_CASE__ = AutoBackbone.from_pretrained(__lowerCamelCase , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip('''TimmBackbone doesn\'t support feed forward chunking''' ) def lowercase_ ( self : str ) -> int: pass @unittest.skip('''TimmBackbone doesn\'t have num_hidden_layers attribute''' ) def lowercase_ ( self : Union[str, Any] ) -> int: pass @unittest.skip('''TimmBackbone initialization is managed on the timm side''' ) def lowercase_ ( self : Any ) -> Dict: pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def lowercase_ ( self : int ) -> Dict: pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def lowercase_ ( self : List[Any] ) -> Optional[Any]: pass @unittest.skip('''TimmBackbone model cannot be created without specifying a backbone checkpoint''' ) def lowercase_ ( self : Optional[Any] ) -> int: pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def lowercase_ ( self : Any ) -> List[str]: pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def lowercase_ ( self : Dict ) -> str: pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def lowercase_ ( self : Optional[Any] ) -> int: pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def lowercase_ ( self : Dict ) -> List[Any]: pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def lowercase_ ( self : List[Any] ) -> str: pass @unittest.skip('''TimmBackbone doesn\'t have hidden size info in its configuration.''' ) def lowercase_ ( self : List[str] ) -> List[Any]: pass @unittest.skip('''TimmBackbone doesn\'t support output_attentions.''' ) def lowercase_ ( self : List[str] ) -> Optional[Any]: pass @unittest.skip('''Safetensors is not supported by timm.''' ) def lowercase_ ( self : Tuple ) -> Optional[int]: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def lowercase_ ( self : Any ) -> List[Any]: pass def lowercase_ ( self : Dict ) -> List[str]: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def lowercase_ ( self : Dict ) -> Optional[int]: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.has_attentions # no need to test all models as different heads yield the same functionality SCREAMING_SNAKE_CASE__ = self.all_model_classes[0] SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = outputs[0][-1] # Encoder-/Decoder-only models SCREAMING_SNAKE_CASE__ = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: SCREAMING_SNAKE_CASE__ = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=__lowerCamelCase ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def lowercase_ ( self : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None SCREAMING_SNAKE_CASE__ = copy.deepcopy(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights SCREAMING_SNAKE_CASE__ = copy.deepcopy(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase )
363
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Optional[Any] = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "vivit" def __init__( self : str , __lowerCamelCase : List[Any]=224 , __lowerCamelCase : Optional[int]=32 , __lowerCamelCase : Tuple=[2, 16, 16] , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : Optional[Any]=768 , __lowerCamelCase : Any=12 , __lowerCamelCase : Optional[Any]=12 , __lowerCamelCase : List[str]=3072 , __lowerCamelCase : Any="gelu_fast" , __lowerCamelCase : Union[str, Any]=0.0 , __lowerCamelCase : int=0.0 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Any=1e-06 , __lowerCamelCase : Dict=True , **__lowerCamelCase : Any , ) -> List[str]: SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = num_frames SCREAMING_SNAKE_CASE__ = tubelet_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = qkv_bias super().__init__(**__lowerCamelCase )
218
0
from __future__ import annotations snake_case_ : Tuple = 10 def A (__A : list[int] ) -> list[int]: """simple docstring""" UpperCAmelCase_ = 1 UpperCAmelCase_ = max(__A ) while placement <= max_digit: # declare and initialize empty buckets UpperCAmelCase_ = [[] for _ in range(__A )] # split list_of_ints between the buckets for i in list_of_ints: UpperCAmelCase_ = int((i / placement) % RADIX ) buckets[tmp].append(__A ) # put each buckets' contents into list_of_ints UpperCAmelCase_ = 0 for b in range(__A ): for i in buckets[b]: UpperCAmelCase_ = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
51
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class __a (UpperCamelCase_): '''simple docstring''' _SCREAMING_SNAKE_CASE :Union[List[PIL.Image.Image], np.ndarray] _SCREAMING_SNAKE_CASE :Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
132
0
"""simple docstring""" import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def __UpperCAmelCase ( snake_case_ : Dict ) -> Union[str, Any]: # picklable for multiprocessing """simple docstring""" return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def __UpperCAmelCase ( ) -> Optional[Any]: """simple docstring""" with parallel_backend("""spark""" ): assert ParallelBackendConfig.backend_name == "spark" _lowerCAmelCase = [1, 2, 3] with pytest.raises(snake_case_ ): with parallel_backend("""unsupported backend""" ): map_nested(snake_case_ , snake_case_ , num_proc=2 ) with pytest.raises(snake_case_ ): with parallel_backend("""unsupported backend""" ): map_nested(snake_case_ , snake_case_ , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize("""num_proc""" , [2, -1] ) def __UpperCAmelCase ( snake_case_ : int ) -> Optional[int]: """simple docstring""" _lowerCAmelCase = [1, 2] _lowerCAmelCase = {"""a""": 1, """b""": 2} _lowerCAmelCase = {"""a""": [1, 2], """b""": [3, 4]} _lowerCAmelCase = {"""a""": {"""1""": 1}, """b""": 2} _lowerCAmelCase = {"""a""": 1, """b""": 2, """c""": 3, """d""": 4} _lowerCAmelCase = [2, 3] _lowerCAmelCase = {"""a""": 2, """b""": 3} _lowerCAmelCase = {"""a""": [2, 3], """b""": [4, 5]} _lowerCAmelCase = {"""a""": {"""1""": 2}, """b""": 3} _lowerCAmelCase = {"""a""": 2, """b""": 3, """c""": 4, """d""": 5} with parallel_backend("""spark""" ): assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa
317
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import SeqaSeqTrainer from seqaseq_training_args import SeqaSeqTrainingArguments import transformers from transformers import ( AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer, HfArgumentParser, MBartTokenizer, MBartTokenizerFast, set_seed, ) from transformers.trainer_utils import EvaluationStrategy, is_main_process from transformers.training_args import ParallelMode from utils import ( SeqaSeqDataCollator, SeqaSeqDataset, assert_all_frozen, build_compute_metrics_fn, check_output_dir, freeze_embeds, freeze_params, lmap, save_json, use_task_specific_params, write_txt_file, ) SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) @dataclass class __lowerCamelCase : __UpperCamelCase = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __UpperCamelCase = field( default=__lowercase , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) __UpperCamelCase = field( default=__lowercase , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) __UpperCamelCase = field( default=__lowercase , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) __UpperCamelCase = field(default=__lowercase , metadata={'help': 'Whether tp freeze the encoder.'} ) __UpperCamelCase = field(default=__lowercase , metadata={'help': 'Whether to freeze the embeddings.'} ) @dataclass class __lowerCamelCase : __UpperCamelCase = field( metadata={'help': 'The input data dir. Should contain the .tsv files (or other data files) for the task.'} ) __UpperCamelCase = field( default='summarization' , metadata={'help': 'Task name, summarization (or summarization_{dataset} for pegasus) or translation'} , ) __UpperCamelCase = field( default=1_024 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __UpperCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total sequence length for target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __UpperCamelCase = field( default=142 , metadata={ 'help': ( 'The maximum total sequence length for validation target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded. ' 'This argument is also used to override the ``max_length`` param of ``model.generate``, which is used ' 'during ``evaluate`` and ``predict``.' ) } , ) __UpperCamelCase = field( default=142 , metadata={ 'help': ( 'The maximum total sequence length for test target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __UpperCamelCase = field(default=-1 , metadata={'help': '# training examples. -1 means use all.'} ) __UpperCamelCase = field(default=-1 , metadata={'help': '# validation examples. -1 means use all.'} ) __UpperCamelCase = field(default=-1 , metadata={'help': '# test examples. -1 means use all.'} ) __UpperCamelCase = field(default=__lowercase , metadata={'help': 'Source language id for translation.'} ) __UpperCamelCase = field(default=__lowercase , metadata={'help': 'Target language id for translation.'} ) __UpperCamelCase = field(default=__lowercase , metadata={'help': '# num_beams to use for evaluation.'} ) __UpperCamelCase = field( default=__lowercase , metadata={'help': 'If only pad tokens should be ignored. This assumes that `config.pad_token_id` is defined.'} , ) def __UpperCAmelCase ( snake_case_ : Optional[int] , snake_case_ : Any , snake_case_ : Union[str, Any] ) -> Tuple: """simple docstring""" logger.info(F"""***** {split} metrics *****""" ) for key in sorted(metrics.keys() ): logger.info(F""" {key} = {metrics[key]}""" ) save_json(snake_case_ , os.path.join(snake_case_ , F"""{split}_results.json""" ) ) def __UpperCAmelCase ( ) -> Union[str, Any]: """simple docstring""" _lowerCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, SeqaSeqTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = parser.parse_args_into_dataclasses() check_output_dir(snake_case_ ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.parallel_mode == ParallelMode.DISTRIBUTED ) , training_args.fpaa , ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("""Training/evaluation parameters %s""" , snake_case_ ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowerCAmelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _lowerCAmelCase = ("""encoder_layerdrop""", """decoder_layerdrop""", """dropout""", """attention_dropout""") for p in extra_model_params: if getattr(snake_case_ , snake_case_ , snake_case_ ): assert hasattr(snake_case_ , snake_case_ ), F"""({config.__class__.__name__}) doesn't have a `{p}` attribute""" setattr(snake_case_ , snake_case_ , getattr(snake_case_ , snake_case_ ) ) _lowerCAmelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _lowerCAmelCase = AutoModelForSeqaSeqLM.from_pretrained( model_args.model_name_or_path , from_tf=""".ckpt""" in model_args.model_name_or_path , config=snake_case_ , cache_dir=model_args.cache_dir , ) # use task specific params use_task_specific_params(snake_case_ , data_args.task ) # set num_beams for evaluation if data_args.eval_beams is None: _lowerCAmelCase = model.config.num_beams # set decoder_start_token_id for MBart if model.config.decoder_start_token_id is None and isinstance(snake_case_ , (MBartTokenizer, MBartTokenizerFast) ): assert ( data_args.tgt_lang is not None and data_args.src_lang is not None ), "mBart requires --tgt_lang and --src_lang" if isinstance(snake_case_ , snake_case_ ): _lowerCAmelCase = tokenizer.lang_code_to_id[data_args.tgt_lang] else: _lowerCAmelCase = tokenizer.convert_tokens_to_ids(data_args.tgt_lang ) if model_args.freeze_embeds: freeze_embeds(snake_case_ ) if model_args.freeze_encoder: freeze_params(model.get_encoder() ) assert_all_frozen(model.get_encoder() ) _lowerCAmelCase = SeqaSeqDataset # Get datasets _lowerCAmelCase = ( dataset_class( snake_case_ , type_path="""train""" , data_dir=data_args.data_dir , n_obs=data_args.n_train , max_target_length=data_args.max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_train else None ) _lowerCAmelCase = ( dataset_class( snake_case_ , type_path="""val""" , data_dir=data_args.data_dir , n_obs=data_args.n_val , max_target_length=data_args.val_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO else None ) _lowerCAmelCase = ( dataset_class( snake_case_ , type_path="""test""" , data_dir=data_args.data_dir , n_obs=data_args.n_test , max_target_length=data_args.test_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_predict else None ) # Initialize our Trainer _lowerCAmelCase = ( build_compute_metrics_fn(data_args.task , snake_case_ ) if training_args.predict_with_generate else None ) _lowerCAmelCase = SeqaSeqTrainer( model=snake_case_ , args=snake_case_ , data_args=snake_case_ , train_dataset=snake_case_ , eval_dataset=snake_case_ , data_collator=SeqaSeqDataCollator( snake_case_ , snake_case_ , model.config.decoder_start_token_id , training_args.tpu_num_cores ) , compute_metrics=snake_case_ , tokenizer=snake_case_ , ) _lowerCAmelCase = {} # Training if training_args.do_train: logger.info("""*** Train ***""" ) _lowerCAmelCase = trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) _lowerCAmelCase = train_result.metrics _lowerCAmelCase = data_args.n_train trainer.save_model() # this also saves the tokenizer if trainer.is_world_process_zero(): handle_metrics("""train""" , snake_case_ , training_args.output_dir ) all_metrics.update(snake_case_ ) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir , """trainer_state.json""" ) ) # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) tokenizer.save_pretrained(training_args.output_dir ) # Evaluation if training_args.do_eval: logger.info("""*** Evaluate ***""" ) _lowerCAmelCase = trainer.evaluate(metric_key_prefix="""val""" ) _lowerCAmelCase = data_args.n_val _lowerCAmelCase = round(metrics["""val_loss"""] , 4 ) if trainer.is_world_process_zero(): handle_metrics("""val""" , snake_case_ , training_args.output_dir ) all_metrics.update(snake_case_ ) if training_args.do_predict: logger.info("""*** Predict ***""" ) _lowerCAmelCase = trainer.predict(test_dataset=snake_case_ , metric_key_prefix="""test""" ) _lowerCAmelCase = test_output.metrics _lowerCAmelCase = data_args.n_test if trainer.is_world_process_zero(): _lowerCAmelCase = round(metrics["""test_loss"""] , 4 ) handle_metrics("""test""" , snake_case_ , training_args.output_dir ) all_metrics.update(snake_case_ ) if training_args.predict_with_generate: _lowerCAmelCase = tokenizer.batch_decode( test_output.predictions , skip_special_tokens=snake_case_ , clean_up_tokenization_spaces=snake_case_ ) _lowerCAmelCase = lmap(str.strip , snake_case_ ) write_txt_file(snake_case_ , os.path.join(training_args.output_dir , """test_generations.txt""" ) ) if trainer.is_world_process_zero(): save_json(snake_case_ , os.path.join(training_args.output_dir , """all_results.json""" ) ) return all_metrics def __UpperCAmelCase ( snake_case_ : Any ) -> Dict: """simple docstring""" main() if __name__ == "__main__": main()
317
1
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Any = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """bart""" UpperCAmelCase__ = ["""past_key_values"""] UpperCAmelCase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Tuple , UpperCAmelCase : Optional[int]=50265 , UpperCAmelCase : Optional[int]=1024 , UpperCAmelCase : Optional[int]=12 , UpperCAmelCase : Dict=4096 , UpperCAmelCase : List[str]=16 , UpperCAmelCase : List[Any]=12 , UpperCAmelCase : List[Any]=4096 , UpperCAmelCase : Optional[Any]=16 , UpperCAmelCase : int=0.0 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : Optional[int]="gelu" , UpperCAmelCase : Tuple=1024 , UpperCAmelCase : int=0.1 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : str=0.0 , UpperCAmelCase : Tuple=0.0_2 , UpperCAmelCase : List[str]=0.0 , UpperCAmelCase : Any=False , UpperCAmelCase : List[str]=True , UpperCAmelCase : str=3 , UpperCAmelCase : Union[str, Any]=1 , UpperCAmelCase : List[str]=0 , UpperCAmelCase : List[Any]=2 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=2 , UpperCAmelCase : List[str]=2 , **UpperCAmelCase : Dict , ) -> Union[str, Any]: lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Tuple = max_position_embeddings lowerCamelCase__ : Optional[Any] = d_model lowerCamelCase__ : int = encoder_ffn_dim lowerCamelCase__ : int = encoder_layers lowerCamelCase__ : Optional[Any] = encoder_attention_heads lowerCamelCase__ : Dict = decoder_ffn_dim lowerCamelCase__ : Union[str, Any] = decoder_layers lowerCamelCase__ : List[Any] = decoder_attention_heads lowerCamelCase__ : List[str] = dropout lowerCamelCase__ : List[Any] = attention_dropout lowerCamelCase__ : Optional[Any] = activation_dropout lowerCamelCase__ : str = activation_function lowerCamelCase__ : List[str] = init_std lowerCamelCase__ : Optional[Any] = encoder_layerdrop lowerCamelCase__ : Optional[int] = decoder_layerdrop lowerCamelCase__ : int = classifier_dropout lowerCamelCase__ : Union[str, Any] = use_cache lowerCamelCase__ : str = encoder_layers lowerCamelCase__ : str = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=UpperCAmelCase , pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , is_encoder_decoder=UpperCAmelCase , decoder_start_token_id=UpperCAmelCase , forced_eos_token_id=UpperCAmelCase , **UpperCAmelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , UpperCAmelCase ): lowerCamelCase__ : Dict = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ 'The config can simply be saved and uploaded again to be fixed.' ) class lowerCAmelCase ( __UpperCamelCase ): @property def A_ ( self : Dict ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Any = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ : Tuple = {0: 'batch'} lowerCamelCase__ : Tuple = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: lowerCamelCase__ : Optional[Any] = {0: 'batch', 1: 'decoder_sequence'} lowerCamelCase__ : Dict = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(UpperCAmelCase , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCamelCase__ : int = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.num_layers for i in range(UpperCAmelCase ): lowerCamelCase__ : List[Any] = {0: 'batch', 2: 'past_sequence + sequence'} lowerCamelCase__ : Optional[Any] = {0: 'batch', 2: 'past_sequence + sequence'} else: lowerCamelCase__ : Union[str, Any] = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def A_ ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : int = super().outputs else: lowerCamelCase__ : Optional[Any] = super(UpperCAmelCase , self ).outputs if self.use_past: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.num_layers for i in range(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = {0: 'batch', 2: 'past_sequence + sequence'} lowerCamelCase__ : List[str] = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def A_ ( self : Optional[Any] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowerCamelCase__ : List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # Generate decoder inputs lowerCamelCase__ : List[str] = seq_length if not self.use_past else 1 lowerCamelCase__ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : List[str] = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} lowerCamelCase__ : Optional[int] = dict(**UpperCAmelCase , **UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ , lowerCamelCase__ : Optional[int] = common_inputs['input_ids'].shape lowerCamelCase__ : List[str] = common_inputs['decoder_input_ids'].shape[1] lowerCamelCase__ , lowerCamelCase__ : List[str] = self.num_attention_heads lowerCamelCase__ : Any = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Dict = decoder_seq_length + 3 lowerCamelCase__ : int = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCamelCase__ : Dict = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(UpperCAmelCase , UpperCAmelCase )] , dim=1 ) lowerCamelCase__ : Any = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCamelCase__ , lowerCamelCase__ : Tuple = self.num_layers lowerCamelCase__ : List[Any] = min(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Tuple = max(UpperCAmelCase , UpperCAmelCase ) - min_num_layers lowerCamelCase__ : str = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(UpperCAmelCase ): common_inputs["past_key_values"].append( ( torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), ) ) # TODO: test this. lowerCamelCase__ : List[str] = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(UpperCAmelCase , UpperCAmelCase ): common_inputs["past_key_values"].append((torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase )) ) return common_inputs def A_ ( self : Tuple , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowerCamelCase__ : Any = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ , lowerCamelCase__ : int = common_inputs['input_ids'].shape # Not using the same length for past_key_values lowerCamelCase__ : Union[str, Any] = seqlen + 2 lowerCamelCase__ , lowerCamelCase__ : List[Any] = self.num_layers lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.num_attention_heads lowerCamelCase__ : Tuple = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Union[str, Any] = common_inputs['attention_mask'].dtype lowerCamelCase__ : int = torch.cat( [common_inputs['attention_mask'], torch.ones(UpperCAmelCase , UpperCAmelCase , dtype=UpperCAmelCase )] , dim=1 ) lowerCamelCase__ : Any = [ (torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase )) for _ in range(UpperCAmelCase ) ] return common_inputs def A_ ( self : Optional[int] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowerCamelCase__ : Tuple = compute_effective_axis_dimension( UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCamelCase__ : List[str] = tokenizer.num_special_tokens_to_add(UpperCAmelCase ) lowerCamelCase__ : List[str] = compute_effective_axis_dimension( UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=UpperCAmelCase ) # Generate dummy inputs according to compute batch and sequence lowerCamelCase__ : str = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCamelCase__ : Dict = dict(tokenizer(UpperCAmelCase , return_tensors=UpperCAmelCase ) ) return common_inputs def A_ ( self : List[Any] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : List[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) elif self.task == "causal-lm": lowerCamelCase__ : int = self._generate_dummy_inputs_for_causal_lm( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) else: lowerCamelCase__ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) return common_inputs def A_ ( self : List[str] , UpperCAmelCase : Tuple , UpperCAmelCase : Any , UpperCAmelCase : Any , UpperCAmelCase : int ) -> List[Any]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Dict = super()._flatten_past_key_values_(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) else: lowerCamelCase__ : str = super(UpperCAmelCase , self )._flatten_past_key_values_( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )
50
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = (PNDMScheduler,) snake_case__ = (("num_inference_steps", 50),) def __lowerCAmelCase ( self : List[str] ,**lowerCamelCase__ : str ): UpperCAmelCase__ = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', } config.update(**lowerCamelCase__ ) return config def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Optional[Any]=0 ,**lowerCamelCase__ : List[str] ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config(**lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals UpperCAmelCase__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class.from_pretrained(lowerCamelCase__ ) new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __lowerCAmelCase ( self : Tuple ): pass def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : List[str]=0 ,**lowerCamelCase__ : Tuple ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class.from_pretrained(lowerCamelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __lowerCAmelCase ( self : List[Any] ,**lowerCamelCase__ : int ): UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(**lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) UpperCAmelCase__ = 10 UpperCAmelCase__ = self.dummy_model() UpperCAmelCase__ = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase__ = model(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase__ = model(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample return sample def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample if num_inference_steps is not None and hasattr(lowerCamelCase__ ,'set_timesteps' ): scheduler.set_timesteps(lowerCamelCase__ ) elif num_inference_steps is not None and not hasattr(lowerCamelCase__ ,'set_timesteps' ): UpperCAmelCase__ = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,0 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,1 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,0 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,1 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def __lowerCAmelCase ( self : List[Any] ): for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowerCamelCase__ ) UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps ,torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) ,) def __lowerCAmelCase ( self : Dict ): for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1] ,[0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=lowerCamelCase__ ,beta_end=lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): for t in [1, 5, 10]: self.check_over_forward(time_step=lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): for t, num_inference_steps in zip([1, 5, 10] ,[10, 50, 100] ): self.check_over_forward(num_inference_steps=lowerCamelCase__ ) def __lowerCAmelCase ( self : int ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase__ = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample def __lowerCAmelCase ( self : int ): with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.step_plms(self.dummy_sample ,1 ,self.dummy_sample ).prev_sample def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.full_loop() UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 1_9_8.1_3_1_8 ) < 1e-2 assert abs(result_mean.item() - 0.2_5_8_0 ) < 1e-3 def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.full_loop(prediction_type='v_prediction' ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 6_7.3_9_8_6 ) < 1e-2 assert abs(result_mean.item() - 0.0_8_7_8 ) < 1e-3 def __lowerCAmelCase ( self : Union[str, Any] ): # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase__ = self.full_loop(set_alpha_to_one=lowerCamelCase__ ,beta_start=0.0_1 ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 2_3_0.0_3_9_9 ) < 1e-2 assert abs(result_mean.item() - 0.2_9_9_5 ) < 1e-3 def __lowerCAmelCase ( self : Tuple ): # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase__ = self.full_loop(set_alpha_to_one=lowerCamelCase__ ,beta_start=0.0_1 ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 1_8_6.9_4_8_2 ) < 1e-2 assert abs(result_mean.item() - 0.2_4_3_4 ) < 1e-3
98
0
"""simple docstring""" import numpy as np def SCREAMING_SNAKE_CASE ( _lowerCamelCase : np.ndarray ,_lowerCamelCase : float ) -> np.ndarray: return np.where(vector > 0 ,_lowerCamelCase ,(alpha * (np.exp(_lowerCamelCase ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
126
"""simple docstring""" import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __A : def __init__( self , a__ , a__=2 , a__=3 , a__=4 , a__=2 , a__=7 , a__=True , a__=True , a__=True , a__=True , a__=99 , a__=36 , a__=3 , a__=4 , a__=37 , a__="gelu" , a__=0.1 , a__=0.1 , a__=512 , a__=16 , a__=2 , a__=0.0_2 , a__=6 , a__=6 , a__=3 , a__=4 , a__=None , a__=1000 , ): _lowerCAmelCase : Union[str, Any] = parent _lowerCAmelCase : Optional[Any] = batch_size _lowerCAmelCase : List[Any] = num_channels _lowerCAmelCase : str = image_size _lowerCAmelCase : str = patch_size _lowerCAmelCase : str = text_seq_length _lowerCAmelCase : List[Any] = is_training _lowerCAmelCase : List[str] = use_input_mask _lowerCAmelCase : List[Any] = use_token_type_ids _lowerCAmelCase : Optional[Any] = use_labels _lowerCAmelCase : Any = vocab_size _lowerCAmelCase : str = hidden_size _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : List[Any] = num_attention_heads _lowerCAmelCase : str = intermediate_size _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Optional[Any] = hidden_dropout_prob _lowerCAmelCase : Any = attention_probs_dropout_prob _lowerCAmelCase : Tuple = max_position_embeddings _lowerCAmelCase : Dict = type_vocab_size _lowerCAmelCase : Tuple = type_sequence_label_size _lowerCAmelCase : Union[str, Any] = initializer_range _lowerCAmelCase : Dict = coordinate_size _lowerCAmelCase : Optional[int] = shape_size _lowerCAmelCase : str = num_labels _lowerCAmelCase : Optional[Any] = num_choices _lowerCAmelCase : str = scope _lowerCAmelCase : Dict = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _lowerCAmelCase : Optional[int] = text_seq_length _lowerCAmelCase : Any = (image_size // patch_size) ** 2 + 1 _lowerCAmelCase : Any = self.text_seq_length + self.image_seq_length def __A ( self ): _lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _lowerCAmelCase : Optional[Any] = bbox[i, j, 3] _lowerCAmelCase : List[str] = bbox[i, j, 1] _lowerCAmelCase : List[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: _lowerCAmelCase : int = bbox[i, j, 2] _lowerCAmelCase : Optional[int] = bbox[i, j, 0] _lowerCAmelCase : Optional[int] = t _lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : List[str] = None if self.use_input_mask: _lowerCAmelCase : int = random_attention_mask([self.batch_size, self.text_seq_length] ) _lowerCAmelCase : str = None if self.use_token_type_ids: _lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _lowerCAmelCase : int = None _lowerCAmelCase : int = None if self.use_labels: _lowerCAmelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : str = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _lowerCAmelCase : str = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : int = LayoutLMvaModel(config=a__ ) model.to(a__ ) model.eval() # text + image _lowerCAmelCase : Optional[Any] = model(a__ , pixel_values=a__ ) _lowerCAmelCase : Any = model( a__ , bbox=a__ , pixel_values=a__ , attention_mask=a__ , token_type_ids=a__ ) _lowerCAmelCase : List[Any] = model(a__ , bbox=a__ , pixel_values=a__ , token_type_ids=a__ ) _lowerCAmelCase : Tuple = model(a__ , bbox=a__ , pixel_values=a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _lowerCAmelCase : Dict = model(a__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _lowerCAmelCase : Optional[Any] = model(pixel_values=a__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : int = self.num_labels _lowerCAmelCase : int = LayoutLMvaForSequenceClassification(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Tuple = model( a__ , bbox=a__ , pixel_values=a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Optional[Any] = self.num_labels _lowerCAmelCase : str = LayoutLMvaForTokenClassification(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : List[str] = model( a__ , bbox=a__ , pixel_values=a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Dict = LayoutLMvaForQuestionAnswering(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : str = model( a__ , bbox=a__ , pixel_values=a__ , attention_mask=a__ , token_type_ids=a__ , start_positions=a__ , end_positions=a__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __A ( self ): _lowerCAmelCase : List[str] = self.prepare_config_and_inputs() ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) : Any = config_and_inputs _lowerCAmelCase : Union[str, Any] = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class __A ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Tuple = False _UpperCamelCase : Any = False _UpperCamelCase : Any = False _UpperCamelCase : int = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase : Union[str, Any] = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def __A ( self , a__ , a__ , a__ , a__ , a__ ): # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def __A ( self ): _lowerCAmelCase : Any = LayoutLMvaModelTester(self ) _lowerCAmelCase : int = ConfigTester(self , config_class=a__ , hidden_size=37 ) def __A ( self , a__ , a__ , a__=False ): _lowerCAmelCase : List[str] = copy.deepcopy(a__ ) if model_class in get_values(a__ ): _lowerCAmelCase : Optional[int] = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(a__ , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(a__ ): _lowerCAmelCase : List[Any] = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=a__ ) elif model_class in get_values(a__ ): _lowerCAmelCase : Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a__ ) _lowerCAmelCase : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a__ ) elif model_class in [ *get_values(a__ ), ]: _lowerCAmelCase : Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a__ ) elif model_class in [ *get_values(a__ ), ]: _lowerCAmelCase : int = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=a__ , ) return inputs_dict def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCAmelCase : int = type self.model_tester.create_and_check_model(*a__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a__ ) def __A ( self ): _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a__ ) def __A ( self ): _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a__ ) @slow def __A ( self ): for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : str = LayoutLMvaModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def SCREAMING_SNAKE_CASE ( ) -> str: _lowerCAmelCase : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class __A ( unittest.TestCase ): @cached_property def __A ( self ): return LayoutLMvaImageProcessor(apply_ocr=a__ ) if is_vision_available() else None @slow def __A ( self ): _lowerCAmelCase : Optional[int] = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(a__ ) _lowerCAmelCase : Dict = self.default_image_processor _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=a__ , return_tensors="""pt""" ).pixel_values.to(a__ ) _lowerCAmelCase : Optional[Any] = torch.tensor([[1, 2]] ) _lowerCAmelCase : Dict = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _lowerCAmelCase : str = model( input_ids=input_ids.to(a__ ) , bbox=bbox.to(a__ ) , pixel_values=pixel_values.to(a__ ) , ) # verify the logits _lowerCAmelCase : Optional[int] = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , a__ ) _lowerCAmelCase : Union[str, Any] = torch.tensor( [[-0.0_5_2_9, 0.3_6_1_8, 0.1_6_3_2], [-0.1_5_8_7, -0.1_6_6_7, -0.0_4_0_0], [-0.1_5_5_7, -0.1_6_7_1, -0.0_5_0_5]] ).to(a__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , a__ , atol=1e-4 ) )
126
1
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
49
import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): # Initialise PyTorch model __a = BigBirdConfig.from_json_file(_UpperCAmelCase ) print(f'Building PyTorch model from configuration: {config}' ) if is_trivia_qa: __a = BigBirdForQuestionAnswering(_UpperCAmelCase ) else: __a = BigBirdForPreTraining(_UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(_UpperCAmelCase , _UpperCAmelCase , is_trivia_qa=_UpperCAmelCase ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __snake_case :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--big_bird_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_trivia_qa''', action='''store_true''', help='''Whether to convert a model with a trivia_qa head.''' ) __snake_case :Any = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
49
1
def lowerCamelCase_ ( UpperCamelCase__ : list ): '''simple docstring''' if not isinstance(UpperCamelCase__, UpperCamelCase__ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(UpperCamelCase__ ) == 0: raise ValueError('''Input list must be a non empty list''' ) if len(UpperCamelCase__ ) == 1: return True UpperCamelCase__ = series[1] - series[0] for index in range(len(UpperCamelCase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def lowerCamelCase_ ( UpperCamelCase__ : list ): '''simple docstring''' if not isinstance(UpperCamelCase__, UpperCamelCase__ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(UpperCamelCase__ ) == 0: raise ValueError('''Input list must be a non empty list''' ) UpperCamelCase__ = 0 for val in series: answer += val return answer / len(UpperCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys lowercase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
35
1
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) lowerCamelCase : int = Vector() def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(__A ) , "(0,0,0,0,0,1)" ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(__A ) , 4 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Union[str, Any] = Vector([1, 2] ) lowerCamelCase : Dict = Vector([1, 2, 3, 4, 5] ) lowerCamelCase : str = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) lowerCamelCase : List[str] = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Tuple = Vector([1, 2, 3] ) lowerCamelCase : Dict = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = Vector([1, 2, 3] ) lowerCamelCase : Optional[int] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : int = Vector([1, 2, 3] ) lowerCamelCase : Union[str, Any] = Vector([2, -1, 4] ) # for test of dot product lowerCamelCase : Optional[Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def _snake_case ( self ): """simple docstring""" self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def _snake_case ( self ): """simple docstring""" self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = Vector([1, 2, 3] ) lowerCamelCase : List[str] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , __A , __A ) ) , "(3,4,7)" ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) lowerCamelCase : List[str] = x.copy() self.assertEqual(str(__A ) , str(__A ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : int = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(__A ) , "(0,1,0)" ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(__A ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCamelCase : int = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(__A , __A ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCamelCase : int = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(__A , __A ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : int = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) lowerCamelCase : int = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(__A ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCamelCase : Any = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCamelCase : Optional[int] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def _snake_case ( self ): """simple docstring""" self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
283
from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) _snake_case = 2_99_79_24_58 # Symbols _snake_case , _snake_case , _snake_case , _snake_case = symbols('''ct x y z''') def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if velocity > c: raise ValueError("Speed must not exceed light speed 299,792,458 [m/s]!" ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError("Speed must be greater than or equal to 1!" ) return velocity / c def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return 1 / sqrt(1 - beta(SCREAMING_SNAKE_CASE_ ) ** 2 ) def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return np.array( [ [gamma(SCREAMING_SNAKE_CASE_ ), -gamma(SCREAMING_SNAKE_CASE_ ) * beta(SCREAMING_SNAKE_CASE_ ), 0, 0], [-gamma(SCREAMING_SNAKE_CASE_ ) * beta(SCREAMING_SNAKE_CASE_ ), gamma(SCREAMING_SNAKE_CASE_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None ): '''simple docstring''' if event is None: lowerCamelCase : Tuple = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(SCREAMING_SNAKE_CASE_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: _snake_case = transform(29_97_92_45) print('''Example of four vector: ''') print(f'''ct\' = {four_vector[0]}''') print(f'''x\' = {four_vector[1]}''') print(f'''y\' = {four_vector[2]}''') print(f'''z\' = {four_vector[3]}''') # Substitute symbols with numerical values _snake_case = {ct: c, x: 1, y: 1, z: 1} _snake_case = [four_vector[i].subs(sub_dict) for i in range(4)] print(f'''\n{numerical_vector}''')
283
1
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version SCREAMING_SNAKE_CASE__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize SCREAMING_SNAKE_CASE__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' SCREAMING_SNAKE_CASE__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' SCREAMING_SNAKE_CASE__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def A__ ( self ) -> Union[str, Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"""] , reference_urls=[ """https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score""", """https://en.wikipedia.org/wiki/METEOR""", ] , ) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" import nltk nltk.download("""wordnet""" ) if NLTK_VERSION >= version.Version("""3.6.5""" ): nltk.download("""punkt""" ) if NLTK_VERSION >= version.Version("""3.6.6""" ): nltk.download("""omw-1.4""" ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0.9 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=0.5 ) -> List[Any]: """simple docstring""" if NLTK_VERSION >= version.Version("""3.6.5""" ): UpperCamelCase = [ meteor_score.single_meteor_score( word_tokenize(_SCREAMING_SNAKE_CASE ) , word_tokenize(_SCREAMING_SNAKE_CASE ) , alpha=_SCREAMING_SNAKE_CASE , beta=_SCREAMING_SNAKE_CASE , gamma=_SCREAMING_SNAKE_CASE ) for ref, pred in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] else: UpperCamelCase = [ meteor_score.single_meteor_score(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , alpha=_SCREAMING_SNAKE_CASE , beta=_SCREAMING_SNAKE_CASE , gamma=_SCREAMING_SNAKE_CASE ) for ref, pred in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] return {"meteor": np.mean(_SCREAMING_SNAKE_CASE )}
354
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a_ ( lowerCamelCase ): lowercase = """Salesforce/blip-image-captioning-base""" lowercase = ( """This is a tool that generates a description of an image. It takes an input named `image` which should be the """ """image to caption, and returns a text that contains the description in English.""" ) lowercase = """image_captioner""" lowercase = AutoModelForVisionaSeq lowercase = ["""image"""] lowercase = ["""text"""] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" requires_backends(self , ["""vision"""] ) super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" return self.pre_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" return self.model.generate(**_SCREAMING_SNAKE_CASE ) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" return self.pre_processor.batch_decode(_SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE )[0].strip()
183
0
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( A_ , A_ ): while second != 0: lowerCAmelCase__ : Tuple = first & second first ^= second lowerCAmelCase__ : Union[str, Any] = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Optional[Any] = int(input('''Enter the first number: ''').strip()) __UpperCamelCase : List[Any] = int(input('''Enter the second number: ''').strip()) print(F'''{add(first, second) = }''')
106
'''simple docstring''' from torch import nn class lowercase ( nn.Module ): """simple docstring""" def __init__( self ,a_ ,a_ ) -> List[Any]: super().__init__() _UpperCAmelCase : Dict = class_size _UpperCAmelCase : Union[str, Any] = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) _UpperCAmelCase : List[Any] = nn.Linear(a_ ,a_ ) def _snake_case ( self ,a_ ) -> Tuple: # hidden_state = nn.functional.relu(self.mlp1(hidden_state)) # hidden_state = self.mlp2(hidden_state) _UpperCAmelCase : Optional[int] = self.mlp(a_ ) return logits
215
0
from math import asin, atan, cos, radians, sin, sqrt, tan A__ = 6378137.0 A__ = 6356752.314245 A__ = 637_8137 def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> float: """simple docstring""" snake_case__ : int = (AXIS_A - AXIS_B) / AXIS_A snake_case__ : Dict = atan((1 - flattening) * tan(radians(__lowerCAmelCase ) ) ) snake_case__ : List[Any] = atan((1 - flattening) * tan(radians(__lowerCAmelCase ) ) ) snake_case__ : int = radians(__lowerCAmelCase ) snake_case__ : List[Any] = radians(__lowerCAmelCase ) # Equation snake_case__ : Union[str, Any] = sin((phi_a - phi_a) / 2 ) snake_case__ : int = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda snake_case__ : Union[str, Any] = sqrt(sin_sq_phi + (cos(__lowerCAmelCase ) * cos(__lowerCAmelCase ) * sin_sq_lambda) ) return 2 * RADIUS * asin(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
351
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( '''The `inpainting.py` script is outdated. Please use directly `from diffusers import''' ''' StableDiffusionInpaintPipeline` instead.''' )
44
0
import math import sys import cva import numpy as np def a__ ( __UpperCamelCase , __UpperCamelCase ): # For applying gaussian function for each element in matrix. SCREAMING_SNAKE_CASE_ = math.sqrt(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = 1 / (sigma * math.sqrt(2 * math.pi )) return cons * np.exp(-((img / sigma) ** 2) * 0.5 ) def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = kernel_size // 2 return img[x - half : x + half + 1, y - half : y + half + 1] def a__ ( __UpperCamelCase , __UpperCamelCase ): # Creates a gaussian kernel of given dimension. SCREAMING_SNAKE_CASE_ = np.zeros((kernel_size, kernel_size) ) for i in range(0 , __UpperCamelCase ): for j in range(0 , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = math.sqrt( abs(i - kernel_size // 2 ) ** 2 + abs(j - kernel_size // 2 ) ** 2 ) return vec_gaussian(__UpperCamelCase , __UpperCamelCase ) def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): SCREAMING_SNAKE_CASE_ = np.zeros(img.shape ) SCREAMING_SNAKE_CASE_ = get_gauss_kernel(__UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = img.shape for i in range(kernel_size // 2 , size_x - kernel_size // 2 ): for j in range(kernel_size // 2 , size_y - kernel_size // 2 ): SCREAMING_SNAKE_CASE_ = get_slice(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ = img_s - img_s[kernel_size // 2, kernel_size // 2] SCREAMING_SNAKE_CASE_ = vec_gaussian(__UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ = np.multiply(__UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ = np.multiply(__UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ = np.sum(__UpperCamelCase ) / np.sum(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = val return imga def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = args[1] if args[1:] else "../image_data/lena.jpg" SCREAMING_SNAKE_CASE_ = float(args[2] ) if args[2:] else 1.0 SCREAMING_SNAKE_CASE_ = float(args[3] ) if args[3:] else 1.0 if args[4:]: SCREAMING_SNAKE_CASE_ = int(args[4] ) SCREAMING_SNAKE_CASE_ = kernel_size + abs(kernel_size % 2 - 1 ) else: SCREAMING_SNAKE_CASE_ = 5 return filename, spatial_variance, intensity_variance, kernel_size if __name__ == "__main__": A , A , A , A : Optional[int] = parse_args(sys.argv) A : Optional[Any] = cva.imread(filename, 0) cva.imshow("input image", img) A : Optional[int] = img / 2_55 A : List[Any] = out.astype("float32") A : Optional[int] = bilateral_filter(out, spatial_variance, intensity_variance, kernel_size) A : int = out * 2_55 A : str = np.uinta(out) cva.imshow("output image", out) cva.waitKey(0) cva.destroyAllWindows()
118
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification A : Tuple = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co A : Dict = "main" # Default branch name A : List[str] = "f2c752cfc5c0ab6f4bdec59acea69eefbee381c2" # One particular commit (not the top of `main`) A : Tuple = "aaaaaaa" # This commit does not exist, so we should 404. A : int = "d9e9f15bc825e4b2c9249e9578f884bbcb5e3684" # Sha-1 of config.json on the top of `main`, for checking purposes A : Tuple = "4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3" @contextlib.contextmanager def a__ ( ): print("Welcome!" ) yield print("Bye!" ) @contextlib.contextmanager def a__ ( ): print("Bonjour!" ) yield print("Au revoir!" ) class lowerCamelCase (unittest.TestCase ): """simple docstring""" def __A ( self : Union[str, Any] ) -> Any: # If the spec is missing, importlib would not be able to import the module dynamically. assert transformers.__spec__ is not None assert importlib.util.find_spec("transformers" ) is not None class lowerCamelCase (unittest.TestCase ): """simple docstring""" @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO ) def __A ( self : Tuple , __magic_name__ : Union[str, Any] ) -> Union[str, Any]: with ContextManagers([] ): print("Transformers are awesome!" ) # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , "Transformers are awesome!\n" ) @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO ) def __A ( self : Dict , __magic_name__ : Union[str, Any] ) -> int: with ContextManagers([context_en()] ): print("Transformers are awesome!" ) # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , "Welcome!\nTransformers are awesome!\nBye!\n" ) @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO ) def __A ( self : Tuple , __magic_name__ : str ) -> Union[str, Any]: with ContextManagers([context_fr(), context_en()] ): print("Transformers are awesome!" ) # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , "Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n" ) @require_torch def __A ( self : List[str] ) -> Union[str, Any]: self.assertEqual(find_labels(__magic_name__ ) , ["labels"] ) self.assertEqual(find_labels(__magic_name__ ) , ["labels", "next_sentence_label"] ) self.assertEqual(find_labels(__magic_name__ ) , ["start_positions", "end_positions"] ) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" pass self.assertEqual(find_labels(__magic_name__ ) , ["labels"] ) @require_tf def __A ( self : List[str] ) -> Optional[Any]: self.assertEqual(find_labels(__magic_name__ ) , ["labels"] ) self.assertEqual(find_labels(__magic_name__ ) , ["labels", "next_sentence_label"] ) self.assertEqual(find_labels(__magic_name__ ) , ["start_positions", "end_positions"] ) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" pass self.assertEqual(find_labels(__magic_name__ ) , ["labels"] ) @require_flax def __A ( self : int ) -> Tuple: # Flax models don't have labels self.assertEqual(find_labels(__magic_name__ ) , [] ) self.assertEqual(find_labels(__magic_name__ ) , [] ) self.assertEqual(find_labels(__magic_name__ ) , [] ) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" pass self.assertEqual(find_labels(__magic_name__ ) , [] )
118
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: a : Dict = None a : int = logging.get_logger(__name__) a : Tuple = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} a : Tuple = { 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json' ), }, } a : Tuple = { 'moussaKam/mbarthez': 1_024, 'moussaKam/barthez': 1_024, 'moussaKam/barthez-orangesum-title': 1_024, } a : List[str] = '▁' class _a ( _lowerCAmelCase ): A = VOCAB_FILES_NAMES A = PRETRAINED_VOCAB_FILES_MAP A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A = ['''input_ids''', '''attention_mask'''] A = BarthezTokenizer def __init__(self, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_="<s>", SCREAMING_SNAKE_CASE_="</s>", SCREAMING_SNAKE_CASE_="</s>", SCREAMING_SNAKE_CASE_="<s>", SCREAMING_SNAKE_CASE_="<unk>", SCREAMING_SNAKE_CASE_="<pad>", SCREAMING_SNAKE_CASE_="<mask>", **SCREAMING_SNAKE_CASE_, ) -> Dict: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_: str = AddedToken(SCREAMING_SNAKE_CASE_, lstrip=SCREAMING_SNAKE_CASE_, rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) else mask_token super().__init__( SCREAMING_SNAKE_CASE_, tokenizer_file=SCREAMING_SNAKE_CASE_, bos_token=SCREAMING_SNAKE_CASE_, eos_token=SCREAMING_SNAKE_CASE_, unk_token=SCREAMING_SNAKE_CASE_, sep_token=SCREAMING_SNAKE_CASE_, cls_token=SCREAMING_SNAKE_CASE_, pad_token=SCREAMING_SNAKE_CASE_, mask_token=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_, ) UpperCAmelCase_: str = vocab_file UpperCAmelCase_: Union[str, Any] = False if not self.vocab_file else True def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_: Union[str, Any] = [self.cls_token_id] UpperCAmelCase_: int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None ) -> List[int]: UpperCAmelCase_: List[Any] = [self.sep_token_id] UpperCAmelCase_: List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return UpperCAmelCase_: str = os.path.join( SCREAMING_SNAKE_CASE_, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE_ ): copyfile(self.vocab_file, SCREAMING_SNAKE_CASE_ ) return (out_vocab_file,)
82
def lowerCAmelCase_ (lowerCAmelCase__: int , lowerCAmelCase__: float , lowerCAmelCase__: float ): """simple docstring""" return round(float(moles / volume ) * nfactor ) def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: float , lowerCAmelCase__: float ): """simple docstring""" return round(float((moles * 0.0821 * temperature) / (volume) ) ) def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: float , lowerCAmelCase__: float ): """simple docstring""" return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: float , lowerCAmelCase__: float ): """simple docstring""" return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
82
1
from collections.abc import Iterable from typing import Generic, TypeVar _UpperCAmelCase : int =TypeVar("""_T""") class snake_case__( Generic[_T] ): '''simple docstring''' def __init__( self , __lowercase = None ) -> None: lowerCAmelCase_ : list[_T] = list(iterable or [] ) lowerCAmelCase_ : list[_T] = [] def __len__( self ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def lowercase_ ( self , __lowercase ) -> None: self._stacka.append(__lowercase ) def lowercase_ ( self ) -> _T: lowerCAmelCase_ : int = self._stacka.pop lowerCAmelCase_ : List[str] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError('''Queue is empty''' ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
"""simple docstring""" import torch def _a ( ): """simple docstring""" if torch.cuda.is_available(): UpperCAmelCase = torch.cuda.device_count() else: UpperCAmelCase = 0 print(F'''Successfully ran on {num_gpus} GPUs''' ) if __name__ == "__main__": main()
234
"""simple docstring""" _UpperCamelCase = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCamelCase = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCamelCase = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" assert len(str(_snake_case ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: UpperCAmelCase = year // 100 UpperCAmelCase = (5 * (century % 4) + 2) % 7 UpperCAmelCase = year % 100 UpperCAmelCase = centurian % 12 UpperCAmelCase = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 UpperCAmelCase = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) UpperCAmelCase = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
234
1
"""simple docstring""" from __future__ import annotations import math def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> int: if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if not scores: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) ) def __UpperCAmelCase ( ) -> None: lowercase__ : Any = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] lowercase__ : Optional[Any] = math.log(len(__lowerCamelCase ) , 2 ) print(f"""Optimal value : {minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
16
'''simple docstring''' import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import logging lowerCamelCase :Any = logging.get_logger(__name__) # pylint: disable=invalid-name class _lowerCAmelCase ( __UpperCAmelCase ): def __init__(self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ): super().__init__() self.register_modules( vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , unet=lowercase , scheduler=lowercase , safety_checker=lowercase , feature_extractor=lowercase , ) def _a (self , lowercase = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory A_ : int = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase ) def _a (self ): self.enable_attention_slicing(lowercase ) @torch.no_grad() def __call__(self , lowercase , lowercase = 512 , lowercase = 512 , lowercase = 50 , lowercase = 7.5 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = None , lowercase = None , lowercase = "pil" , lowercase = True , lowercase = None , lowercase = 1 , lowercase = None , **lowercase , ): if isinstance(lowercase , lowercase ): A_ : Union[str, Any] = 1 elif isinstance(lowercase , lowercase ): A_ : Any = len(lowercase ) else: raise ValueError(F'`prompt` has to be of type `str` or `list` but is {type(lowercase )}' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F'`height` and `width` have to be divisible by 8 but are {height} and {width}.' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowercase , lowercase ) or callback_steps <= 0) ): raise ValueError( F'`callback_steps` has to be a positive integer but is {callback_steps} of type' F' {type(lowercase )}.' ) # get prompt text embeddings A_ : Optional[Any] = self.tokenizer( lowercase , padding="""max_length""" , max_length=self.tokenizer.model_max_length , return_tensors="""pt""" , ) A_ : Dict = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: A_ : Union[str, Any] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F' {self.tokenizer.model_max_length} tokens: {removed_text}' ) A_ : Dict = text_input_ids[:, : self.tokenizer.model_max_length] if text_embeddings is None: A_ : Any = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method A_, A_, A_ : Tuple = text_embeddings.shape A_ : Optional[Any] = text_embeddings.repeat(1 , lowercase , 1 ) A_ : Any = text_embeddings.view(bs_embed * num_images_per_prompt , lowercase , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. A_ : List[str] = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: A_ : List[str] if negative_prompt is None: A_ : Optional[int] = [""""""] elif type(lowercase ) is not type(lowercase ): raise TypeError( F'`negative_prompt` should be the same type to `prompt`, but got {type(lowercase )} !=' F' {type(lowercase )}.' ) elif isinstance(lowercase , lowercase ): A_ : Dict = [negative_prompt] elif batch_size != len(lowercase ): raise ValueError( F'`negative_prompt`: {negative_prompt} has batch size {len(lowercase )}, but `prompt`:' F' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches' """ the batch size of `prompt`.""" ) else: A_ : Dict = negative_prompt A_ : int = text_input_ids.shape[-1] A_ : List[Any] = self.tokenizer( lowercase , padding="""max_length""" , max_length=lowercase , truncation=lowercase , return_tensors="""pt""" , ) A_ : str = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method A_ : Optional[Any] = uncond_embeddings.shape[1] A_ : str = uncond_embeddings.repeat(lowercase , lowercase , 1 ) A_ : List[str] = uncond_embeddings.view(batch_size * num_images_per_prompt , lowercase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes A_ : Dict = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. A_ : int = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) A_ : Dict = (batch_size * num_images_per_prompt, self.unet.config.in_channels, 64, 64) A_ : Dict = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps A_ : Tuple = torch.randn( lowercase , generator=lowercase , device="""cpu""" , dtype=lowercase ).to(self.device ) A_ : int = torch.randn(lowercase , generator=lowercase , device="""cpu""" , dtype=lowercase ).to( self.device ) else: A_ : int = torch.randn( lowercase , generator=lowercase , device=self.device , dtype=lowercase ) A_ : str = torch.randn(lowercase , generator=lowercase , device=self.device , dtype=lowercase ) else: if latents_reference.shape != latents_shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {latents_shape}' ) A_ : str = latents_reference.to(self.device ) A_ : Tuple = latents.to(self.device ) # This is the key part of the pipeline where we # try to ensure that the generated images w/ the same seed # but different sizes actually result in similar images A_ : Optional[int] = (latents_shape[3] - latents_shape_reference[3]) // 2 A_ : Optional[int] = (latents_shape[2] - latents_shape_reference[2]) // 2 A_ : Optional[int] = latents_shape_reference[3] if dx >= 0 else latents_shape_reference[3] + 2 * dx A_ : int = latents_shape_reference[2] if dy >= 0 else latents_shape_reference[2] + 2 * dy A_ : Optional[Any] = 0 if dx < 0 else dx A_ : Optional[Any] = 0 if dy < 0 else dy A_ : Optional[int] = max(-dx , 0 ) A_ : List[str] = max(-dy , 0 ) # import pdb # pdb.set_trace() A_ : str = latents_reference[:, :, dy : dy + h, dx : dx + w] # set timesteps self.scheduler.set_timesteps(lowercase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand A_ : Any = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler A_ : Tuple = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] A_ : Dict = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) A_ : Any = {} if accepts_eta: A_ : Optional[int] = eta for i, t in enumerate(self.progress_bar(lowercase ) ): # expand the latents if we are doing classifier free guidance A_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents A_ : Tuple = self.scheduler.scale_model_input(lowercase , lowercase ) # predict the noise residual A_ : List[str] = self.unet(lowercase , lowercase , encoder_hidden_states=lowercase ).sample # perform guidance if do_classifier_free_guidance: A_, A_ : str = noise_pred.chunk(2 ) A_ : Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 A_ : List[str] = self.scheduler.step(lowercase , lowercase , lowercase , **lowercase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowercase , lowercase , lowercase ) A_ : List[str] = 1 / 0.1_82_15 * latents A_ : List[str] = self.vae.decode(lowercase ).sample A_ : Optional[int] = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 A_ : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if self.safety_checker is not None: A_ : Union[str, Any] = self.feature_extractor(self.numpy_to_pil(lowercase ) , return_tensors="""pt""" ).to( self.device ) A_, A_ : Optional[int] = self.safety_checker( images=lowercase , clip_input=safety_checker_input.pixel_values.to(text_embeddings.dtype ) ) else: A_ : Tuple = None if output_type == "pil": A_ : Tuple = self.numpy_to_pil(lowercase ) if not return_dict: return (image, has_nsfw_concept) return StableDiffusionPipelineOutput(images=lowercase , nsfw_content_detected=lowercase )
206
0
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast lowercase__ : int = datasets.utils.logging.get_logger(__name__) @dataclass class UpperCamelCase__ ( datasets.BuilderConfig ): """simple docstring""" _SCREAMING_SNAKE_CASE = 1_0000 _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None class UpperCamelCase__ ( datasets.ArrowBasedBuilder ): """simple docstring""" _SCREAMING_SNAKE_CASE = ParquetConfig def SCREAMING_SNAKE_CASE__ ( self : Dict ): return datasets.DatasetInfo(features=self.config.features ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[str] ): if not self.config.data_files: raise ValueError(F"At least one data file must be specified, but got data_files={self.config.data_files}" ) lowerCAmelCase_ : List[str] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(SCREAMING_SNAKE_CASE_ , (str, list, tuple) ): lowerCAmelCase_ : Optional[int] = data_files if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase_ : Dict = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive lowerCAmelCase_ : List[str] = [dl_manager.iter_files(SCREAMING_SNAKE_CASE_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files} )] lowerCAmelCase_ : Optional[int] = [] for split_name, files in data_files.items(): if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase_ : Any = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive lowerCAmelCase_ : Tuple = [dl_manager.iter_files(SCREAMING_SNAKE_CASE_ ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(SCREAMING_SNAKE_CASE_ ): with open(SCREAMING_SNAKE_CASE_ , 'rb' ) as f: lowerCAmelCase_ : List[str] = datasets.Features.from_arrow_schema(pq.read_schema(SCREAMING_SNAKE_CASE_ ) ) break splits.append(datasets.SplitGenerator(name=SCREAMING_SNAKE_CASE_ , gen_kwargs={'files': files} ) ) return splits def SCREAMING_SNAKE_CASE__ ( self : Dict , SCREAMING_SNAKE_CASE_ : pa.Table ): if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase_ : int = table_cast(SCREAMING_SNAKE_CASE_ , self.info.features.arrow_schema ) return pa_table def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] ): lowerCAmelCase_ : str = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F"Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'" ) for file_idx, file in enumerate(itertools.chain.from_iterable(SCREAMING_SNAKE_CASE_ ) ): with open(SCREAMING_SNAKE_CASE_ , 'rb' ) as f: lowerCAmelCase_ : Optional[Any] = pq.ParquetFile(SCREAMING_SNAKE_CASE_ ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): lowerCAmelCase_ : Optional[int] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F"{file_idx}_{batch_idx}", self._cast_table(SCREAMING_SNAKE_CASE_ ) except ValueError as e: logger.error(F"Failed to read file '{file}' with error {type(SCREAMING_SNAKE_CASE_ )}: {e}" ) raise
289
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : str ) -> List[Any]: """simple docstring""" return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : str=0 ) -> Union[str, Any]: """simple docstring""" return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : x[column] ) def UpperCamelCase_ ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any]=float('inf' ) ) -> Optional[int]: """simple docstring""" for i in range(points_counts - 1 ): for j in range(i + 1 , lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: lowerCAmelCase_ : Optional[int] = current_dis return min_dis def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Union[str, Any]=float('inf' ) ) -> Dict: """simple docstring""" for i in range(min(6 , points_counts - 1 ) , lowerCAmelCase__ ): for j in range(max(0 , i - 6 ) , lowerCAmelCase__ ): lowerCAmelCase_ : Any = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: lowerCAmelCase_ : Union[str, Any] = current_dis return min_dis def UpperCamelCase_ ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : int , lowerCAmelCase__ : Optional[int] ) -> Dict: """simple docstring""" if points_counts <= 3: return dis_between_closest_pair(lowerCAmelCase__ , lowerCAmelCase__ ) # recursion lowerCAmelCase_ : int = points_counts // 2 lowerCAmelCase_ : Optional[Any] = closest_pair_of_points_sqr( lowerCAmelCase__ , points_sorted_on_y[:mid] , lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = closest_pair_of_points_sqr( lowerCAmelCase__ , points_sorted_on_y[mid:] , points_counts - mid ) lowerCAmelCase_ : Any = min(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase_ : str = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = dis_between_closest_in_strip( lowerCAmelCase__ , len(lowerCAmelCase__ ) , lowerCAmelCase__ ) return min(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : Tuple ) -> List[Any]: """simple docstring""" lowerCAmelCase_ : List[str] = column_based_sort(lowerCAmelCase__ , column=0 ) lowerCAmelCase_ : Dict = column_based_sort(lowerCAmelCase__ , column=1 ) return ( closest_pair_of_points_sqr( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) ) ** 0.5 if __name__ == "__main__": lowercase__ : List[str] = [(2, 3), (1_2, 3_0), (4_0, 5_0), (5, 1), (1_2, 1_0), (3, 4)] print("""Distance:""", closest_pair_of_points(points, len(points)))
289
1
import itertools import string from collections.abc import Generator, Iterable def a_ ( __lowercase : Iterable[str] , __lowercase : int ) -> Generator[tuple[str, ...], None, None]: _snake_case = iter(__lowercase ) while True: _snake_case = tuple(itertools.islice(__lowercase , __lowercase ) ) if not chunk: return yield chunk def a_ ( __lowercase : str ) -> str: _snake_case = ''.join([c.upper() for c in dirty if c in string.ascii_letters] ) _snake_case = '' if len(__lowercase ) < 2: return dirty for i in range(len(__lowercase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(__lowercase ) & 1: clean += "X" return clean def a_ ( __lowercase : str ) -> list[str]: # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) _snake_case = 'ABCDEFGHIKLMNOPQRSTUVWXYZ' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _snake_case = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(__lowercase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(__lowercase ) return table def a_ ( __lowercase : str , __lowercase : str ) -> str: _snake_case = generate_table(__lowercase ) _snake_case = prepare_input(__lowercase ) _snake_case = '' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowercase , 2 ): _snake_case , _snake_case = divmod(table.index(__lowercase ) , 5 ) _snake_case , _snake_case = divmod(table.index(__lowercase ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def a_ ( __lowercase : str , __lowercase : str ) -> str: _snake_case = generate_table(__lowercase ) _snake_case = '' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowercase , 2 ): _snake_case , _snake_case = divmod(table.index(__lowercase ) , 5 ) _snake_case , _snake_case = divmod(table.index(__lowercase ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
282
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCamelCase : List[Any] = HfApi() _lowerCamelCase : Dict = {} # fmt: off _lowerCamelCase : List[Any] = torch.tensor([ -0.7_5_1_5, -1.6_8_8_3, 0.2_4_2_0, 0.0_3_0_0, 0.6_3_4_7, 1.3_4_3_3, -1.1_7_4_3, -3.7_4_6_7, 1.2_3_4_2, -2.2_4_8_5, 0.4_6_3_6, 0.8_0_7_6, -0.7_9_9_1, 0.3_9_6_9, 0.8_4_9_8, 0.9_1_8_9, -1.8_8_8_7, -3.3_5_2_2, 0.7_6_3_9, 0.2_0_4_0, 0.6_2_7_1, -2.7_1_4_8, -1.6_3_1_6, 3.0_8_3_9, 0.3_1_8_6, 0.2_7_2_1, -0.9_7_5_9, -1.2_4_6_1, 2.6_2_5_7, 1.3_5_5_7 ]) _lowerCamelCase : int = torch.tensor([ -2.3_6_3_9, -2.5_3_4_4, 0.0_0_5_4, -0.6_6_7_4, 1.5_9_9_0, 1.0_1_5_8, 0.3_1_2_4, -2.1_4_3_6, 1.8_7_9_5, -2.5_4_2_9, -0.1_5_6_6, -0.3_9_7_3, 1.2_4_9_0, 2.6_4_4_7, 1.2_2_8_3, -0.5_2_0_8, -2.8_1_5_4, -3.5_1_1_9, 2.3_8_3_8, 1.2_0_3_3, 1.7_2_0_1, -2.1_2_5_6, -1.4_5_7_6, 2.7_9_4_8, 2.4_2_0_4, -0.9_7_5_2, -1.2_5_4_6, 0.8_0_2_7, 3.2_7_5_8, 3.1_3_6_5 ]) _lowerCamelCase : Optional[int] = torch.tensor([ -0.6_5_3_1, -0.6_8_9_1, -0.3_1_7_2, -0.5_3_7_5, -0.9_1_4_0, -0.5_3_6_7, -0.1_1_7_5, -0.7_8_6_9, -0.3_8_0_8, -0.4_5_1_3, -0.2_0_9_8, -0.0_0_8_3, 0.3_1_8_3, 0.5_1_4_0, 0.2_2_4_7, -0.1_3_0_4, -0.1_3_0_2, -0.2_8_0_2, -0.2_0_8_4, -0.2_0_2_5, -0.4_9_6_7, -0.4_8_7_3, -0.0_8_6_1, 0.6_9_2_5, 0.0_2_5_0, 0.1_2_9_0, -0.1_5_4_3, 0.6_3_1_6, 1.0_4_6_0, 1.4_9_4_3 ]) _lowerCamelCase : Dict = torch.tensor([ 0.0_9_1_1, 0.1_1_0_7, 0.0_1_8_2, 0.0_4_3_5, -0.0_8_0_5, -0.0_6_0_8, 0.0_3_8_1, 0.2_1_7_2, -0.0_2_8_0, 0.1_3_2_7, -0.0_2_9_9, -0.0_2_5_5, -0.0_0_5_0, -0.1_1_7_0, -0.1_0_4_6, 0.0_3_0_9, 0.1_3_6_7, 0.1_7_2_8, -0.0_5_3_3, -0.0_7_4_8, -0.0_5_3_4, 0.1_6_2_4, 0.0_3_8_4, -0.1_8_0_5, -0.0_7_0_7, 0.0_6_4_2, 0.0_2_2_0, -0.0_1_3_4, -0.1_3_3_3, -0.1_5_0_5 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_3_2_1, 0.1_3_3_7, 0.0_4_4_0, 0.0_6_2_2, -0.0_5_9_1, -0.0_3_7_0, 0.0_5_0_3, 0.2_1_3_3, -0.0_1_7_7, 0.1_4_1_5, -0.0_1_1_6, -0.0_1_1_2, 0.0_0_4_4, -0.0_9_8_0, -0.0_7_8_9, 0.0_3_9_5, 0.1_5_0_2, 0.1_7_8_5, -0.0_4_8_8, -0.0_5_1_4, -0.0_4_0_4, 0.1_5_3_9, 0.0_4_5_4, -0.1_5_5_9, -0.0_6_6_5, 0.0_6_5_9, 0.0_3_8_3, -0.0_0_0_5, -0.1_2_6_6, -0.1_3_8_6 ]) _lowerCamelCase : List[Any] = torch.tensor([ 0.1_1_5_4, 0.1_2_1_8, 0.0_3_0_7, 0.0_5_2_6, -0.0_7_1_1, -0.0_5_4_1, 0.0_3_6_6, 0.2_0_7_8, -0.0_2_6_7, 0.1_3_1_7, -0.0_2_2_6, -0.0_1_9_3, -0.0_0_1_4, -0.1_0_5_5, -0.0_9_0_2, 0.0_3_3_0, 0.1_3_9_1, 0.1_7_0_9, -0.0_5_6_2, -0.0_6_9_3, -0.0_5_6_0, 0.1_4_8_2, 0.0_3_8_1, -0.1_6_8_3, -0.0_6_8_1, 0.0_6_6_1, 0.0_3_3_1, -0.0_0_4_6, -0.1_2_6_8, -0.1_4_3_1 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_1_9_2, 0.1_2_4_0, 0.0_4_1_4, 0.0_6_0_6, -0.0_5_5_7, -0.0_4_1_2, 0.0_4_3_0, 0.2_0_4_2, -0.0_2_0_0, 0.1_3_8_5, -0.0_1_1_5, -0.0_1_3_2, 0.0_0_1_7, -0.0_9_6_5, -0.0_8_0_2, 0.0_3_9_8, 0.1_4_3_3, 0.1_7_4_7, -0.0_4_5_8, -0.0_5_3_3, -0.0_4_0_7, 0.1_5_4_5, 0.0_4_1_9, -0.1_5_7_4, -0.0_6_4_5, 0.0_6_2_6, 0.0_3_4_1, -0.0_0_1_0, -0.1_1_9_9, -0.1_3_9_0 ]) _lowerCamelCase : int = torch.tensor([ 0.1_0_7_5, 0.1_0_7_4, 0.0_2_0_5, 0.0_4_3_1, -0.0_7_7_4, -0.0_6_0_7, 0.0_2_9_8, 0.2_0_4_2, -0.0_3_2_0, 0.1_2_6_7, -0.0_2_8_1, -0.0_2_5_0, -0.0_0_6_4, -0.1_0_9_1, -0.0_9_4_6, 0.0_2_9_0, 0.1_3_2_8, 0.1_6_5_0, -0.0_5_8_0, -0.0_7_3_8, -0.0_5_8_6, 0.1_4_4_0, 0.0_3_3_7, -0.1_7_4_6, -0.0_7_1_2, 0.0_6_0_5, 0.0_2_5_0, -0.0_0_9_9, -0.1_3_1_6, -0.1_4_7_3 ]) _lowerCamelCase : int = torch.tensor([ -1.4_5_7_2, -2.0_4_8_1, -0.0_4_1_4, -0.6_0_0_5, 1.4_1_3_6, 0.5_8_4_8, 0.4_0_2_8, -2.7_3_3_0, 1.2_2_1_2, -2.1_2_2_8, 0.2_1_5_5, 0.4_0_3_9, 0.7_6_6_2, 2.0_5_3_5, 0.7_4_7_7, -0.3_2_4_3, -2.1_7_5_8, -2.7_6_4_8, 1.6_9_4_7, 0.7_0_2_6, 1.2_3_3_8, -1.6_0_7_8, -0.8_6_8_2, 2.2_8_1_0, 1.8_5_7_4, -0.5_7_1_8, -0.5_5_8_6, -0.0_1_8_6, 2.3_4_1_5, 2.1_2_5_1]) _lowerCamelCase : Tuple = torch.tensor([ -1.3_6_9_0, -1.9_7_2_0, -0.4_0_9_0, -0.6_9_6_6, 1.4_6_6_0, 0.9_9_3_8, -0.1_3_8_5, -2.7_3_2_4, 0.7_7_3_6, -1.8_9_1_7, 0.2_9_2_3, 0.4_2_9_3, 0.1_6_9_3, 1.4_1_1_2, 1.1_8_8_7, -0.3_1_8_1, -2.2_1_6_0, -2.6_3_8_1, 1.3_1_7_0, 0.8_1_6_3, 0.9_2_4_0, -1.6_5_4_4, -0.6_0_9_9, 2.5_2_5_9, 1.6_4_3_0, -0.9_0_9_0, -0.9_3_9_2, -0.0_1_2_6, 2.4_2_6_8, 2.3_2_6_6 ]) _lowerCamelCase : List[str] = torch.tensor([ -1.3_5_2_5, -1.9_6_2_8, -0.3_9_5_6, -0.6_8_6_0, 1.4_6_6_4, 1.0_0_1_4, -0.1_2_5_9, -2.7_2_1_2, 0.7_7_7_2, -1.8_8_1_1, 0.2_9_9_6, 0.4_3_8_8, 0.1_7_0_4, 1.4_0_2_9, 1.1_7_0_1, -0.3_0_2_7, -2.2_0_5_3, -2.6_2_8_7, 1.3_3_5_0, 0.8_1_3_1, 0.9_2_7_4, -1.6_2_9_2, -0.6_0_9_8, 2.5_1_3_1, 1.6_5_0_5, -0.8_9_5_8, -0.9_2_9_8, -0.0_1_5_1, 2.4_2_5_7, 2.3_3_5_5 ]) _lowerCamelCase : int = torch.tensor([ -2.0_5_8_5, -2.7_8_9_7, -0.2_8_5_0, -0.8_9_4_0, 1.9_0_5_2, 0.5_7_0_2, 0.6_3_4_5, -3.8_9_5_9, 1.5_9_3_2, -3.2_3_1_9, 0.1_9_7_4, 0.0_2_8_7, 1.7_5_6_6, 2.6_5_4_3, 0.8_3_8_7, -0.5_3_5_1, -3.2_7_3_6, -4.3_3_7_5, 2.9_0_2_9, 1.6_3_9_0, 1.4_6_4_0, -2.1_7_0_1, -1.9_0_1_3, 2.9_3_4_1, 3.4_9_8_1, -0.6_2_5_5, -1.1_6_4_4, -0.1_5_9_1, 3.7_0_9_7, 3.2_0_6_6 ]) _lowerCamelCase : Tuple = torch.tensor([ -2.3_1_3_9, -2.5_5_9_4, -0.0_1_9_7, -0.6_7_8_5, 1.7_0_0_1, 1.1_6_0_6, 0.3_0_7_5, -2.1_7_4_0, 1.8_0_7_1, -2.5_6_3_0, -0.0_9_2_6, -0.3_8_1_1, 1.2_1_1_6, 2.6_2_4_6, 1.2_7_3_1, -0.5_3_9_8, -2.8_1_5_3, -3.6_1_4_0, 2.3_8_9_3, 1.3_2_6_2, 1.6_2_5_8, -2.1_8_5_6, -1.3_2_6_7, 2.8_3_9_5, 2.3_7_7_9, -1.0_6_2_3, -1.2_4_6_8, 0.8_9_5_9, 3.3_3_6_7, 3.2_2_4_3 ]) _lowerCamelCase : int = torch.tensor([ -2.0_6_2_8, -2.7_6_6_7, -0.2_0_8_9, -0.8_2_6_3, 2.0_5_3_9, 0.5_9_9_2, 0.6_4_9_5, -3.8_3_3_6, 1.6_0_2_5, -3.2_8_1_7, 0.1_7_2_1, -0.0_6_3_3, 1.7_5_1_6, 2.7_0_3_9, 0.8_1_0_0, -0.5_9_0_8, -3.2_1_1_3, -4.4_3_4_3, 2.9_2_5_7, 1.3_6_3_2, 1.5_5_6_2, -2.1_4_8_9, -1.9_8_9_4, 3.0_5_6_0, 3.3_3_9_6, -0.7_3_2_8, -1.0_4_1_7, 0.0_3_8_3, 3.7_0_9_3, 3.2_3_4_3 ]) _lowerCamelCase : List[Any] = torch.tensor([ -1.4_5_7_4, -2.0_5_6_9, -0.0_4_7_3, -0.6_1_1_7, 1.4_0_1_8, 0.5_7_6_9, 0.4_1_2_9, -2.7_3_4_4, 1.2_2_4_1, -2.1_3_9_7, 0.2_0_0_0, 0.3_9_3_7, 0.7_6_1_6, 2.0_4_5_3, 0.7_3_2_4, -0.3_3_9_1, -2.1_7_4_6, -2.7_7_4_4, 1.6_9_6_3, 0.6_9_2_1, 1.2_1_8_7, -1.6_1_7_2, -0.8_8_7_7, 2.2_4_3_9, 1.8_4_7_1, -0.5_8_3_9, -0.5_6_0_5, -0.0_4_6_4, 2.3_2_5_0, 2.1_2_1_9 ]) # fmt: on _lowerCamelCase : List[str] = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": _lowerCamelCase : Any = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(F'Started running {mod.modelId}!!!') if mod.modelId.startswith('''CompVis'''): _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: _lowerCamelCase : int = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) _lowerCamelCase : Union[str, Any] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) _lowerCamelCase : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): _lowerCamelCase : int = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1E-3 ) print(F'{mod.modelId} has passed successfully!!!')
282
1
from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _lowerCamelCase : str = logging.get_logger(__name__) def __lowerCamelCase (UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] ): return [ int(1_0_0_0 * (box[0] / width) ), int(1_0_0_0 * (box[1] / height) ), int(1_0_0_0 * (box[2] / width) ), int(1_0_0_0 * (box[3] / height) ), ] def __lowerCamelCase (UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Optional[str] , UpperCAmelCase__ : Optional[str] ): SCREAMING_SNAKE_CASE = to_pil_image(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = pil_image.size SCREAMING_SNAKE_CASE = pytesseract.image_to_data(UpperCAmelCase__ , lang=UpperCAmelCase__ , output_type="dict" , config=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates SCREAMING_SNAKE_CASE = [idx for idx, word in enumerate(UpperCAmelCase__ ) if not word.strip()] SCREAMING_SNAKE_CASE = [word for idx, word in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format SCREAMING_SNAKE_CASE = [] for x, y, w, h in zip(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ): SCREAMING_SNAKE_CASE = [x, y, x + w, y + h] actual_boxes.append(UpperCAmelCase__ ) # finally, normalize the bounding boxes SCREAMING_SNAKE_CASE = [] for box in actual_boxes: normalized_boxes.append(normalize_box(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class lowercase ( a ): lowercase__ : Optional[int] = ["""pixel_values"""] def __init__( self : int , _UpperCamelCase : bool = True , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : PILImageResampling = PILImageResampling.BILINEAR , _UpperCamelCase : bool = True , _UpperCamelCase : float = 1 / 255 , _UpperCamelCase : bool = True , _UpperCamelCase : Union[float, Iterable[float]] = None , _UpperCamelCase : Union[float, Iterable[float]] = None , _UpperCamelCase : bool = True , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[str] = "" , **_UpperCamelCase : Union[str, Any] , ) -> None: '''simple docstring''' super().__init__(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = size if size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = resample SCREAMING_SNAKE_CASE = do_rescale SCREAMING_SNAKE_CASE = rescale_value SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE = image_std if image_std is not None else IMAGENET_STANDARD_STD SCREAMING_SNAKE_CASE = apply_ocr SCREAMING_SNAKE_CASE = ocr_lang SCREAMING_SNAKE_CASE = tesseract_config def __snake_case( self : Dict , _UpperCamelCase : np.ndarray , _UpperCamelCase : Dict[str, int] , _UpperCamelCase : PILImageResampling = PILImageResampling.BILINEAR , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : List[Any] , ) -> np.ndarray: '''simple docstring''' SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) SCREAMING_SNAKE_CASE = (size["height"], size["width"]) return resize(_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : Union[str, Any] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[int, float] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Optional[int] , ) -> np.ndarray: '''simple docstring''' return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : int , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[float, Iterable[float]] , _UpperCamelCase : Union[float, Iterable[float]] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Optional[Any] , ) -> np.ndarray: '''simple docstring''' return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : Tuple , _UpperCamelCase : ImageInput , _UpperCamelCase : bool = None , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : Optional[Any]=None , _UpperCamelCase : bool = None , _UpperCamelCase : float = None , _UpperCamelCase : bool = None , _UpperCamelCase : Union[float, Iterable[float]] = None , _UpperCamelCase : Union[float, Iterable[float]] = None , _UpperCamelCase : bool = None , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[Union[str, TensorType]] = None , _UpperCamelCase : ChannelDimension = ChannelDimension.FIRST , **_UpperCamelCase : List[Any] , ) -> PIL.Image.Image: '''simple docstring''' SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE = size if size is not None else self.size SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE = apply_ocr if apply_ocr is not None else self.apply_ocr SCREAMING_SNAKE_CASE = ocr_lang if ocr_lang is not None else self.ocr_lang SCREAMING_SNAKE_CASE = tesseract_config if tesseract_config is not None else self.tesseract_config SCREAMING_SNAKE_CASE = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("If do_normalize is True, image_mean and image_std must be specified." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(_UpperCamelCase ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , "pytesseract" ) SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for image in images: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = apply_tesseract(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) words_batch.append(_UpperCamelCase ) boxes_batch.append(_UpperCamelCase ) if do_resize: SCREAMING_SNAKE_CASE = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = BatchFeature(data={"pixel_values": images} , tensor_type=_UpperCamelCase ) if apply_ocr: SCREAMING_SNAKE_CASE = words_batch SCREAMING_SNAKE_CASE = boxes_batch return data
366
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( a ): lowercase__ : Dict = (UniPCMultistepScheduler,) lowercase__ : Optional[int] = (("""num_inference_steps""", 25),) def __snake_case( self : List[str] , **_UpperCamelCase : Tuple ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = { "num_train_timesteps": 1_000, "beta_start": 0.0_0_0_1, "beta_end": 0.0_2, "beta_schedule": "linear", "solver_order": 2, "solver_type": "bh2", } config.update(**_UpperCamelCase ) return config def __snake_case( self : List[str] , _UpperCamelCase : Dict=0 , **_UpperCamelCase : List[str] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class.from_pretrained(_UpperCamelCase ) new_scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals SCREAMING_SNAKE_CASE = dummy_past_residuals[: new_scheduler.config.solver_order] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = sample, sample for t in range(_UpperCamelCase , time_step + scheduler.config.solver_order + 1 ): SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = new_scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __snake_case( self : Any , _UpperCamelCase : Union[str, Any]=0 , **_UpperCamelCase : Dict ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals (must be after setting timesteps) SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class.from_pretrained(_UpperCamelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residual (must be after setting timesteps) SCREAMING_SNAKE_CASE = dummy_past_residuals[: new_scheduler.config.solver_order] SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = new_scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __snake_case( self : List[str] , _UpperCamelCase : Tuple=None , **_UpperCamelCase : List[Any] ) -> str: '''simple docstring''' if scheduler is None: SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = 10 SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter scheduler.set_timesteps(_UpperCamelCase ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE = model(_UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ).prev_sample return sample def __snake_case( self : int ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample if num_inference_steps is not None and hasattr(_UpperCamelCase , "set_timesteps" ): scheduler.set_timesteps(_UpperCamelCase ) elif num_inference_steps is not None and not hasattr(_UpperCamelCase , "set_timesteps" ): SCREAMING_SNAKE_CASE = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] SCREAMING_SNAKE_CASE = scheduler.timesteps[5] SCREAMING_SNAKE_CASE = scheduler.timesteps[6] SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __snake_case( self : str ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = UniPCMultistepScheduler(**self.get_scheduler_config() ) SCREAMING_SNAKE_CASE = self.full_loop(scheduler=_UpperCamelCase ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 SCREAMING_SNAKE_CASE = DPMSolverSinglestepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = DEISMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = UniPCMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = self.full_loop(scheduler=_UpperCamelCase ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 def __snake_case( self : Optional[int] ) -> List[Any]: '''simple docstring''' for timesteps in [25, 50, 100, 999, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCamelCase ) def __snake_case( self : Tuple ) -> Union[str, Any]: '''simple docstring''' self.check_over_configs(thresholding=_UpperCamelCase ) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=_UpperCamelCase , prediction_type=_UpperCamelCase , sample_max_value=_UpperCamelCase , solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , ) def __snake_case( self : Tuple ) -> int: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCamelCase ) def __snake_case( self : Dict ) -> int: '''simple docstring''' for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , prediction_type=_UpperCamelCase , ) SCREAMING_SNAKE_CASE = self.full_loop( solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , prediction_type=_UpperCamelCase , ) assert not torch.isnan(_UpperCamelCase ).any(), "Samples have nan numbers" def __snake_case( self : List[str] ) -> Tuple: '''simple docstring''' self.check_over_configs(lower_order_final=_UpperCamelCase ) self.check_over_configs(lower_order_final=_UpperCamelCase ) def __snake_case( self : Optional[int] ) -> List[Any]: '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1_000]: self.check_over_forward(num_inference_steps=_UpperCamelCase , time_step=0 ) def __snake_case( self : Tuple ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = self.full_loop() SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 def __snake_case( self : List[str] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.full_loop(prediction_type="v_prediction" ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.1_0_1_4 ) < 1e-3 def __snake_case( self : int ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(thresholding=_UpperCamelCase , dynamic_thresholding_ratio=0 ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = 10 SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter.half() scheduler.set_timesteps(_UpperCamelCase ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE = model(_UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ).prev_sample assert sample.dtype == torch.floataa def __snake_case( self : List[str] , **_UpperCamelCase : Dict ) -> Union[str, Any]: '''simple docstring''' for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(scheduler.config.num_train_timesteps ) assert len(scheduler.timesteps.unique() ) == scheduler.num_inference_steps
206
0
"""simple docstring""" from datetime import datetime import requests def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[int] = 'https://downloadgram.net/wp-json/wppress/video-downloader/video?url=' A_ : Union[str, Any] = requests.get(base_url + url ).json()[0]['urls'][0]['src'] return requests.get(_snake_case ).content if __name__ == "__main__": lowerCamelCase_ : Dict = input('Enter Video/IGTV url: ').strip() lowerCamelCase_ : Optional[int] = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.mp4" with open(file_name, 'wb') as fp: fp.write(download_video(url)) print(F"Done. Video saved to disk as {file_name}.")
286
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = ['image_processor', 'tokenizer'] SCREAMING_SNAKE_CASE = 'BridgeTowerImageProcessor' SCREAMING_SNAKE_CASE = ('RobertaTokenizer', 'RobertaTokenizerFast') def __init__( self , __snake_case , __snake_case ) -> Optional[int]: '''simple docstring''' super().__init__(__snake_case , __snake_case ) def __call__( self , __snake_case , __snake_case = None , __snake_case = True , __snake_case = False , __snake_case = None , __snake_case = None , __snake_case = 0 , __snake_case = None , __snake_case = None , __snake_case = None , __snake_case = False , __snake_case = False , __snake_case = False , __snake_case = False , __snake_case = True , __snake_case = None , **__snake_case , ) -> BatchEncoding: '''simple docstring''' __a =self.tokenizer( text=__snake_case , add_special_tokens=__snake_case , padding=__snake_case , truncation=__snake_case , max_length=__snake_case , stride=__snake_case , pad_to_multiple_of=__snake_case , return_token_type_ids=__snake_case , return_attention_mask=__snake_case , return_overflowing_tokens=__snake_case , return_special_tokens_mask=__snake_case , return_offsets_mapping=__snake_case , return_length=__snake_case , verbose=__snake_case , return_tensors=__snake_case , **__snake_case , ) # add pixel_values + pixel_mask __a =self.image_processor( __snake_case , return_tensors=__snake_case , do_normalize=__snake_case , do_center_crop=__snake_case , **__snake_case ) encoding.update(__snake_case ) return encoding def __magic_name__ ( self , *__snake_case , **__snake_case ) -> str: '''simple docstring''' return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def __magic_name__ ( self , *__snake_case , **__snake_case ) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__snake_case , **__snake_case ) @property def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' __a =self.tokenizer.model_input_names __a =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
218
0
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Any = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : int = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", ) ) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", ) ) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def SCREAMING_SNAKE_CASE ( snake_case_ : Dict , snake_case_ : Optional[int] , snake_case_ : Tuple ): snake_case__ : Optional[int] = state_dict.pop(snake_case_ ) snake_case__ : List[Any] = val def SCREAMING_SNAKE_CASE ( snake_case_ : Optional[int] ): snake_case__ : List[str] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case__ : List[str] = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) snake_case__ : Optional[int] = value else: snake_case__ : List[str] = value return new_state_dict def SCREAMING_SNAKE_CASE ( snake_case_ : Union[str, Any] ): snake_case__ : List[Any] = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case__ : Union[str, Any] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case__ : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case__ : Dict = in_proj_weight[:256, :] snake_case__ : List[Any] = in_proj_bias[:256] snake_case__ : Union[str, Any] = in_proj_weight[256:512, :] snake_case__ : Optional[Any] = in_proj_bias[256:512] snake_case__ : Any = in_proj_weight[-256:, :] snake_case__ : List[str] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention snake_case__ : Any = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case__ : Tuple = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case__ : List[Any] = in_proj_weight[:256, :] snake_case__ : Any = in_proj_bias[:256] snake_case__ : List[str] = in_proj_weight[256:512, :] snake_case__ : Optional[Any] = in_proj_bias[256:512] snake_case__ : str = in_proj_weight[-256:, :] snake_case__ : Dict = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention snake_case__ : List[Any] = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) snake_case__ : List[Any] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict snake_case__ : str = in_proj_weight_cross_attn[:256, :] snake_case__ : Tuple = in_proj_bias_cross_attn[:256] snake_case__ : Dict = in_proj_weight_cross_attn[256:512, :] snake_case__ : int = in_proj_bias_cross_attn[256:512] snake_case__ : Union[str, Any] = in_proj_weight_cross_attn[-256:, :] snake_case__ : Any = in_proj_bias_cross_attn[-256:] def SCREAMING_SNAKE_CASE ( snake_case_ : Tuple , snake_case_ : List[Any] ): snake_case__, snake_case__ : Tuple = image.size snake_case__ : Optional[Any] = max(snake_case_ , snake_case_ ) snake_case__ : Tuple = 800 if "detection" in checkpoint_url else 1000 snake_case__ : int = target_max_size / current_max_size snake_case__ : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def SCREAMING_SNAKE_CASE ( snake_case_ : Tuple ): snake_case__ : Any = F.to_tensor(snake_case_ ) snake_case__ : Tuple = F.normalize(snake_case_ , mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( snake_case_ : Optional[Any] , snake_case_ : List[str] , snake_case_ : Dict ): logger.info("Converting model..." ) # load original state dict snake_case__ : Union[str, Any] = torch.hub.load_state_dict_from_url(snake_case_ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(snake_case_ , snake_case_ , snake_case_ ) snake_case__ : Tuple = rename_backbone_keys(snake_case_ ) # query, key and value matrices need special treatment read_in_q_k_v(snake_case_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case__ : Union[str, Any] = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case__ : List[Any] = state_dict.pop(snake_case_ ) snake_case__ : Optional[Any] = val # create HuggingFace model and load state dict snake_case__ : Union[str, Any] = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: snake_case__ : Dict = 15 snake_case__ : Dict = 2 snake_case__ : Optional[int] = {0: "table", 1: "table rotated"} snake_case__ : int = idalabel snake_case__ : Optional[int] = {v: k for k, v in idalabel.items()} else: snake_case__ : int = 125 snake_case__ : Tuple = 6 snake_case__ : str = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } snake_case__ : Dict = idalabel snake_case__ : List[str] = {v: k for k, v in idalabel.items()} snake_case__ : Any = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) snake_case__ : Union[str, Any] = TableTransformerForObjectDetection(snake_case_ ) model.load_state_dict(snake_case_ ) model.eval() # verify our conversion snake_case__ : int = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" snake_case__ : List[str] = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=snake_case_ ) snake_case__ : List[str] = Image.open(snake_case_ ).convert("RGB" ) snake_case__ : int = normalize(resize(snake_case_ , snake_case_ ) ).unsqueeze(0 ) snake_case__ : Union[str, Any] = model(snake_case_ ) if "detection" in checkpoint_url: snake_case__ : List[str] = (1, 15, 3) snake_case__ : Union[str, Any] = torch.tensor( [[-6.78_97, -16.99_85, 6.79_37], [-8.01_86, -22.21_92, 6.96_77], [-7.31_17, -21.07_08, 7.40_55]] ) snake_case__ : Optional[Any] = torch.tensor([[0.48_67, 0.17_67, 0.67_32], [0.67_18, 0.44_79, 0.38_30], [0.47_16, 0.17_60, 0.63_64]] ) else: snake_case__ : int = (1, 125, 7) snake_case__ : List[Any] = torch.tensor( [[-18.14_30, -8.32_14, 4.82_74], [-18.46_85, -7.13_61, -4.26_67], [-26.36_93, -9.34_29, -4.99_62]] ) snake_case__ : Dict = torch.tensor([[0.49_83, 0.55_95, 0.94_40], [0.49_16, 0.63_15, 0.59_54], [0.61_08, 0.86_37, 0.11_35]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , snake_case_ , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , snake_case_ , atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(snake_case_ ).mkdir(exist_ok=snake_case_ ) model.save_pretrained(snake_case_ ) image_processor.save_pretrained(snake_case_ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) snake_case__ : Dict = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(snake_case_ ) image_processor.push_to_hub(snake_case_ ) if __name__ == "__main__": __lowerCamelCase : str = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) __lowerCamelCase : Tuple = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
286
__lowerCamelCase : Optional[int] = """Tobias Carryer""" from time import time class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self : List[Any] , __A : List[Any] , __A : Optional[int] , __A : List[str] , __A : Dict=int(time() ) ): # noqa: B008 snake_case__ : List[Any] = multiplier snake_case__ : Optional[int] = increment snake_case__ : Optional[int] = modulo snake_case__ : Union[str, Any] = seed def _lowercase ( self : str ): snake_case__ : Union[str, Any] = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. __lowerCamelCase : int = LinearCongruentialGenerator(166_4525, 10_1390_4223, 2 << 31) while True: print(lcg.next_number())
286
1
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def lowercase ( SCREAMING_SNAKE_CASE__ : Tuple ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def lowercase ( ) -> int: with parallel_backend("""spark""" ): assert ParallelBackendConfig.backend_name == "spark" _snake_case : List[Any] = [1, 2, 3] with pytest.raises(SCREAMING_SNAKE_CASE__ ): with parallel_backend("""unsupported backend""" ): map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=2 ) with pytest.raises(SCREAMING_SNAKE_CASE__ ): with parallel_backend("""unsupported backend""" ): map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize("""num_proc""" , [2, -1] ) def lowercase ( SCREAMING_SNAKE_CASE__ : Tuple ) -> List[str]: _snake_case : int = [1, 2] _snake_case : List[Any] = {"""a""": 1, """b""": 2} _snake_case : Optional[int] = {"""a""": [1, 2], """b""": [3, 4]} _snake_case : int = {"""a""": {"""1""": 1}, """b""": 2} _snake_case : str = {"""a""": 1, """b""": 2, """c""": 3, """d""": 4} _snake_case : Tuple = [2, 3] _snake_case : List[str] = {"""a""": 2, """b""": 3} _snake_case : List[str] = {"""a""": [2, 3], """b""": [4, 5]} _snake_case : Any = {"""a""": {"""1""": 2}, """b""": 3} _snake_case : str = {"""a""": 2, """b""": 3, """c""": 4, """d""": 5} with parallel_backend("""spark""" ): assert map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ ) == expected_map_nested_sa assert map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ ) == expected_map_nested_sa assert map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ ) == expected_map_nested_sa assert map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ ) == expected_map_nested_sa assert map_nested(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ ) == expected_map_nested_sa
317
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) a__ = pytest.mark.integration @pytest.mark.parametrize("""path""" , ["""paws""", """csv"""] ) def lowercase ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Tuple: inspect_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case : Union[str, Any] = path + """.py""" assert script_name in os.listdir(SCREAMING_SNAKE_CASE__ ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE__ ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""" , ["""accuracy"""] ) def lowercase ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Optional[int]: inspect_metric(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case : Dict = path + """.py""" assert script_name in os.listdir(SCREAMING_SNAKE_CASE__ ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( """path, config_name, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[Any]: _snake_case : Dict = get_dataset_config_info(SCREAMING_SNAKE_CASE__ , config_name=SCREAMING_SNAKE_CASE__ ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Tuple: with pytest.raises(SCREAMING_SNAKE_CASE__ ): get_dataset_config_info(SCREAMING_SNAKE_CASE__ , config_name=SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( """path, expected""" , [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int ) -> Optional[Any]: _snake_case : Optional[Any] = get_dataset_config_names(SCREAMING_SNAKE_CASE__ ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""" , [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple ) -> Optional[Any]: _snake_case : Union[str, Any] = get_dataset_infos(SCREAMING_SNAKE_CASE__ ) assert list(infos.keys() ) == expected_configs _snake_case : Optional[int] = expected_configs[0] assert expected_config in infos _snake_case : int = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int ) -> Tuple: _snake_case : Dict = get_dataset_infos(SCREAMING_SNAKE_CASE__ ) assert expected_config in infos _snake_case : Optional[int] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def lowercase ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ) -> Optional[Any]: with pytest.raises(SCREAMING_SNAKE_CASE__ ): get_dataset_split_names(SCREAMING_SNAKE_CASE__ , config_name=SCREAMING_SNAKE_CASE__ )
317
1
import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def lowerCAmelCase_ ( ) -> Any: '''simple docstring''' UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument("--model_ckpt" , type=snake_case_ , default="microsoft/unixcoder-base-nine" ) parser.add_argument("--num_epochs" , type=snake_case_ , default=5 ) parser.add_argument("--batch_size" , type=snake_case_ , default=6 ) parser.add_argument("--gradient_accumulation_steps" , type=snake_case_ , default=1 ) parser.add_argument("--freeze" , type=snake_case_ , default=snake_case_ ) parser.add_argument("--learning_rate" , type=snake_case_ , default=5E-4 ) parser.add_argument("--seed" , type=snake_case_ , default=0 ) parser.add_argument("--lr_scheduler_type" , type=snake_case_ , default="cosine" ) parser.add_argument("--num_warmup_steps" , type=snake_case_ , default=10 ) parser.add_argument("--weight_decay" , type=snake_case_ , default=0.01 ) parser.add_argument("--output_dir" , type=snake_case_ , default="./results" ) return parser.parse_args() SCREAMING_SNAKE_CASE_: List[str] =load('accuracy') def lowerCAmelCase_ ( snake_case_ : List[str] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = eval_pred UpperCAmelCase_ = np.argmax(snake_case_ , axis=1 ) return metric.compute(predictions=snake_case_ , references=snake_case_ ) class __A ( UpperCamelCase__ ): def __init__(self : List[Any] , __a : Any ): super().__init__() UpperCAmelCase_ = trainer def _lowercase (self : Optional[Any] , __a : Union[str, Any] , __a : Union[str, Any] , __a : Optional[int] , **__a : Dict ): if control.should_evaluate: UpperCAmelCase_ = deepcopy(__a ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="train" ) return control_copy def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = get_args() set_seed(args.seed ) UpperCAmelCase_ = load_dataset("codeparrot/codecomplex" , split="train" ) UpperCAmelCase_ = dataset.train_test_split(test_size=0.2 ) UpperCAmelCase_ = train_test["test"].train_test_split(test_size=0.5 ) UpperCAmelCase_ = DatasetDict( { "train": train_test["train"], "test": test_validation["train"], "valid": test_validation["test"], } ) print("Loading tokenizer and model" ) UpperCAmelCase_ = AutoTokenizer.from_pretrained(args.model_ckpt ) UpperCAmelCase_ = tokenizer.eos_token UpperCAmelCase_ = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt , num_labels=7 ) UpperCAmelCase_ = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): UpperCAmelCase_ = False UpperCAmelCase_ = ClassLabel(num_classes=7 , names=list(set(train_test_validation["train"]["complexity"] ) ) ) def tokenize(snake_case_ : Union[str, Any] ): UpperCAmelCase_ = tokenizer(example["src"] , truncation=snake_case_ , max_length=10_24 ) UpperCAmelCase_ = labels.straint(example["complexity"] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } UpperCAmelCase_ = train_test_validation.map( snake_case_ , batched=snake_case_ , remove_columns=train_test_validation["train"].column_names , ) UpperCAmelCase_ = DataCollatorWithPadding(tokenizer=snake_case_ ) UpperCAmelCase_ = TrainingArguments( output_dir=args.output_dir , learning_rate=args.learning_rate , lr_scheduler_type=args.lr_scheduler_type , evaluation_strategy="epoch" , save_strategy="epoch" , logging_strategy="epoch" , per_device_train_batch_size=args.batch_size , per_device_eval_batch_size=args.batch_size , num_train_epochs=args.num_epochs , gradient_accumulation_steps=args.gradient_accumulation_steps , weight_decay=0.01 , metric_for_best_model="accuracy" , run_name="complexity-java" , report_to="wandb" , ) UpperCAmelCase_ = Trainer( model=snake_case_ , args=snake_case_ , train_dataset=tokenized_datasets["train"] , eval_dataset=tokenized_datasets["valid"] , tokenizer=snake_case_ , data_collator=snake_case_ , compute_metrics=snake_case_ , ) print("Training..." ) trainer.add_callback(CustomCallback(snake_case_ ) ) trainer.train() if __name__ == "__main__": main()
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline SCREAMING_SNAKE_CASE_: Dict =logging.get_logger(__name__) class __A ( UpperCamelCase__ ): def _lowercase (self : Dict , __a : Any ): if isinstance(__a , __a ): UpperCAmelCase_ = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__(self : Union[str, Any] , __a : Optional[int] , __a : Optional[int] , __a : int ): if len(__a ) == 0 or len(__a ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(__a ) ) if isinstance(__a , __a ): UpperCAmelCase_ = [sequences] UpperCAmelCase_ = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(__a )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCamelCase__ ) class __A ( UpperCamelCase__ ): def __init__(self : Optional[Any] , __a : Union[str, Any]=ZeroShotClassificationArgumentHandler() , *__a : Optional[int] , **__a : List[str] ): UpperCAmelCase_ = args_parser super().__init__(*__a , **__a ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def _lowercase (self : str ): for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def _lowercase (self : Any , __a : Any , __a : int=True , __a : Dict=True , __a : Any=TruncationStrategy.ONLY_FIRST , **__a : Tuple ): UpperCAmelCase_ = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) UpperCAmelCase_ = self.tokenizer.eos_token try: UpperCAmelCase_ = self.tokenizer( __a , add_special_tokens=__a , return_tensors=__a , padding=__a , truncation=__a , ) except Exception as e: if "too short" in str(__a ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. UpperCAmelCase_ = self.tokenizer( __a , add_special_tokens=__a , return_tensors=__a , padding=__a , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def _lowercase (self : List[str] , **__a : Tuple ): if kwargs.get("multi_class" , __a ) is not None: UpperCAmelCase_ = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) UpperCAmelCase_ = {} if "candidate_labels" in kwargs: UpperCAmelCase_ = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: UpperCAmelCase_ = kwargs["hypothesis_template"] UpperCAmelCase_ = {} if "multi_label" in kwargs: UpperCAmelCase_ = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__(self : Tuple , __a : Union[str, List[str]] , *__a : Optional[Any] , **__a : Tuple , ): if len(__a ) == 0: pass elif len(__a ) == 1 and "candidate_labels" not in kwargs: UpperCAmelCase_ = args[0] else: raise ValueError(f"""Unable to understand extra arguments {args}""" ) return super().__call__(__a , **__a ) def _lowercase (self : Optional[int] , __a : Optional[Any] , __a : List[str]=None , __a : Any="This example is {}." ): UpperCAmelCase_ , UpperCAmelCase_ = self._args_parser(__a , __a , __a ) for i, (candidate_label, sequence_pair) in enumerate(zip(__a , __a ) ): UpperCAmelCase_ = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(__a ) - 1, **model_input, } def _lowercase (self : List[str] , __a : Any ): UpperCAmelCase_ = inputs["candidate_label"] UpperCAmelCase_ = inputs["sequence"] UpperCAmelCase_ = {k: inputs[k] for k in self.tokenizer.model_input_names} UpperCAmelCase_ = self.model(**__a ) UpperCAmelCase_ = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def _lowercase (self : Optional[Any] , __a : List[str] , __a : Tuple=False ): UpperCAmelCase_ = [outputs["candidate_label"] for outputs in model_outputs] UpperCAmelCase_ = [outputs["sequence"] for outputs in model_outputs] UpperCAmelCase_ = np.concatenate([output["logits"].numpy() for output in model_outputs] ) UpperCAmelCase_ = logits.shape[0] UpperCAmelCase_ = len(__a ) UpperCAmelCase_ = N // n UpperCAmelCase_ = logits.reshape((num_sequences, n, -1) ) if multi_label or len(__a ) == 1: # softmax over the entailment vs. contradiction dim for each label independently UpperCAmelCase_ = self.entailment_id UpperCAmelCase_ = -1 if entailment_id == 0 else 0 UpperCAmelCase_ = reshaped_outputs[..., [contradiction_id, entailment_id]] UpperCAmelCase_ = np.exp(__a ) / np.exp(__a ).sum(-1 , keepdims=__a ) UpperCAmelCase_ = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels UpperCAmelCase_ = reshaped_outputs[..., self.entailment_id] UpperCAmelCase_ = np.exp(__a ) / np.exp(__a ).sum(-1 , keepdims=__a ) UpperCAmelCase_ = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
106
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class A_ ( A__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ( """This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.""" """It takes two arguments named `image` which should be the original image, and `label` which should be a text """ """describing the elements what should be identified in the segmentation mask. The tool returns the mask.""" ) SCREAMING_SNAKE_CASE_ = """CIDAS/clipseg-rd64-refined""" SCREAMING_SNAKE_CASE_ = """image_segmenter""" SCREAMING_SNAKE_CASE_ = CLIPSegForImageSegmentation SCREAMING_SNAKE_CASE_ = ["""image""", """text"""] SCREAMING_SNAKE_CASE_ = ["""image"""] def __init__( self :List[str] , *lowerCamelCase_ :Any , **lowerCamelCase_ :str ): """simple docstring""" requires_backends(self , ['vision'] ) super().__init__(*lowerCamelCase_ , **lowerCamelCase_ ) def UpperCAmelCase__ ( self :Optional[int] , lowerCamelCase_ :"Image" , lowerCamelCase_ :str ): """simple docstring""" return self.pre_processor(text=[label] , images=[image] , padding=lowerCamelCase_ , return_tensors='pt' ) def UpperCAmelCase__ ( self :List[str] , lowerCamelCase_ :List[Any] ): """simple docstring""" with torch.no_grad(): lowerCamelCase__ : Tuple =self.model(**lowerCamelCase_ ).logits return logits def UpperCAmelCase__ ( self :str , lowerCamelCase_ :Tuple ): """simple docstring""" lowerCamelCase__ : List[str] =outputs.cpu().detach().numpy() lowerCamelCase__ : Any =0 lowerCamelCase__ : List[Any] =1 return Image.fromarray((array * 255).astype(np.uinta ) )
126
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase = logging.get_logger(__name__) class A_ ( A__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ["""pixel_values"""] def __init__( self :Union[str, Any] , lowerCamelCase_ :bool = True , lowerCamelCase_ :Dict[str, int] = None , lowerCamelCase_ :int = 0.9 , lowerCamelCase_ :PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase_ :bool = True , lowerCamelCase_ :Dict[str, int] = None , lowerCamelCase_ :Union[int, float] = 1 / 255 , lowerCamelCase_ :bool = True , lowerCamelCase_ :bool = True , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , **lowerCamelCase_ :Tuple , ): """simple docstring""" super().__init__(**lowerCamelCase_ ) lowerCamelCase__ : str =size if size is not None else {'shortest_edge': 224} lowerCamelCase__ : List[str] =get_size_dict(lowerCamelCase_ , default_to_square=lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =crop_size if crop_size is not None else {'height': 224, 'width': 224} lowerCamelCase__ : str =get_size_dict(lowerCamelCase_ , param_name='crop_size' ) lowerCamelCase__ : Tuple =do_resize lowerCamelCase__ : List[Any] =size lowerCamelCase__ : List[str] =crop_pct lowerCamelCase__ : Union[str, Any] =resample lowerCamelCase__ : List[str] =do_center_crop lowerCamelCase__ : List[str] =crop_size lowerCamelCase__ : List[Any] =do_rescale lowerCamelCase__ : List[str] =rescale_factor lowerCamelCase__ : Tuple =do_normalize lowerCamelCase__ : int =image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN lowerCamelCase__ : List[Any] =image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCAmelCase__ ( self :Any , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Dict[str, int] , lowerCamelCase_ :Optional[float] = None , lowerCamelCase_ :PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :Any , ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =get_size_dict(lowerCamelCase_ , default_to_square=lowerCamelCase_ ) if "shortest_edge" not in size and ("height" not in size or "width" not in size): raise ValueError(f"""size must contain 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) if crop_pct is not None: if "shortest_edge" in size: lowerCamelCase__ : Optional[int] =int(size['shortest_edge'] / crop_pct ) elif "height" in size and "width" in size: if size["height"] == size["width"]: lowerCamelCase__ : Union[str, Any] =int(size['height'] / crop_pct ) else: lowerCamelCase__ : Any =(int(size['height'] / crop_pct ), int(size['width'] / crop_pct )) else: raise ValueError('Invalid size for resize: {}'.format(lowerCamelCase_ ) ) lowerCamelCase__ : Tuple =get_resize_output_image_size(lowerCamelCase_ , size=lowerCamelCase_ , default_to_square=lowerCamelCase_ ) else: if "shortest_edge" in size: lowerCamelCase__ : str =get_resize_output_image_size(lowerCamelCase_ , size=size['shortest_edge'] , default_to_square=lowerCamelCase_ ) elif "height" in size and "width" in size: lowerCamelCase__ : Union[str, Any] =(size['height'], size['width']) else: raise ValueError('Invalid size for resize: {}'.format(lowerCamelCase_ ) ) return resize(lowerCamelCase_ , size=lowerCamelCase_ , resample=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_ ) def UpperCAmelCase__ ( self :Any , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Dict[str, int] , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :str , ): """simple docstring""" lowerCamelCase__ : Tuple =get_size_dict(lowerCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""size must contain 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(lowerCamelCase_ , size=(size['height'], size['width']) , data_format=lowerCamelCase_ , **lowerCamelCase_ ) def UpperCAmelCase__ ( self :int , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Union[int, float] , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :List[str] , ): """simple docstring""" return rescale(lowerCamelCase_ , scale=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_ ) def UpperCAmelCase__ ( self :List[Any] , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Union[float, List[float]] , lowerCamelCase_ :Union[float, List[float]] , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :Tuple , ): """simple docstring""" return normalize(lowerCamelCase_ , mean=lowerCamelCase_ , std=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_ ) def UpperCAmelCase__ ( self :Any , lowerCamelCase_ :ImageInput , lowerCamelCase_ :bool = None , lowerCamelCase_ :Dict[str, int] = None , lowerCamelCase_ :int = None , lowerCamelCase_ :PILImageResampling = None , lowerCamelCase_ :bool = None , lowerCamelCase_ :Dict[str, int] = None , lowerCamelCase_ :bool = None , lowerCamelCase_ :float = None , lowerCamelCase_ :bool = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[str, TensorType]] = None , lowerCamelCase_ :ChannelDimension = ChannelDimension.FIRST , **lowerCamelCase_ :List[str] , ): """simple docstring""" lowerCamelCase__ : Dict =do_resize if do_resize is not None else self.do_resize lowerCamelCase__ : Union[str, Any] =crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase__ : Tuple =resample if resample is not None else self.resample lowerCamelCase__ : Any =do_center_crop if do_center_crop is not None else self.do_center_crop lowerCamelCase__ : Optional[Any] =do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase__ : Optional[int] =rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase__ : Optional[Any] =do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase__ : List[str] =image_mean if image_mean is not None else self.image_mean lowerCamelCase__ : List[Any] =image_std if image_std is not None else self.image_std lowerCamelCase__ : int =size if size is not None else self.size lowerCamelCase__ : Tuple =get_size_dict(lowerCamelCase_ , default_to_square=lowerCamelCase_ ) lowerCamelCase__ : Dict =crop_size if crop_size is not None else self.crop_size lowerCamelCase__ : str =get_size_dict(lowerCamelCase_ , param_name='crop_size' ) lowerCamelCase__ : Dict =make_list_of_images(lowerCamelCase_ ) if not valid_images(lowerCamelCase_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_pct is None: raise ValueError('Crop_pct must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowerCamelCase__ : List[str] =[to_numpy_array(lowerCamelCase_ ) for image in images] if do_resize: lowerCamelCase__ : Tuple =[self.resize(image=lowerCamelCase_ , size=lowerCamelCase_ , crop_pct=lowerCamelCase_ , resample=lowerCamelCase_ ) for image in images] if do_center_crop: lowerCamelCase__ : Union[str, Any] =[self.center_crop(image=lowerCamelCase_ , size=lowerCamelCase_ ) for image in images] if do_rescale: lowerCamelCase__ : str =[self.rescale(image=lowerCamelCase_ , scale=lowerCamelCase_ ) for image in images] if do_normalize: lowerCamelCase__ : Optional[Any] =[self.normalize(image=lowerCamelCase_ , mean=lowerCamelCase_ , std=lowerCamelCase_ ) for image in images] lowerCamelCase__ : Optional[Any] =[to_channel_dimension_format(lowerCamelCase_ , lowerCamelCase_ ) for image in images] lowerCamelCase__ : List[str] ={'pixel_values': images} return BatchFeature(data=lowerCamelCase_ , tensor_type=lowerCamelCase_ )
126
1
import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( _a : int , _a : str , _a : List[Any] ): # Initialise PyTorch model snake_case_ : List[Any] = LxmertConfig.from_json_file(_a ) print(F'''Building PyTorch model from configuration: {config}''' ) snake_case_ : Optional[Any] = LxmertForPreTraining(_a ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(_a , _a , _a ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , _a ) if __name__ == "__main__": lowercase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowercase : Optional[int] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
36
def lowerCAmelCase__ ( _a : float , _a : float ): if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
36
1
'''simple docstring''' import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class UpperCAmelCase_ ( _a ): """simple docstring""" def __init__( self : Optional[Any] , snake_case_ : NestedDataStructureLike[PathLike] , snake_case_ : Optional[NamedSplit] = None , snake_case_ : Optional[Features] = None , snake_case_ : str = None , snake_case_ : bool = False , snake_case_ : bool = False , snake_case_ : Optional[str] = None , snake_case_ : Optional[int] = None , **snake_case_ : List[Any] , ): super().__init__( snake_case_ , split=snake_case_ , features=snake_case_ , cache_dir=snake_case_ , keep_in_memory=snake_case_ , streaming=snake_case_ , num_proc=snake_case_ , **snake_case_ , ) snake_case__ : Any = field snake_case__ : Optional[Any] = path_or_paths if isinstance(snake_case_ , snake_case_ ) else {self.split: path_or_paths} snake_case__ : Any = Json( cache_dir=snake_case_ , data_files=snake_case_ , features=snake_case_ , field=snake_case_ , **snake_case_ , ) def lowerCamelCase ( self : int ): # Build iterable dataset if self.streaming: snake_case__ : int = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: snake_case__ : Union[str, Any] = None snake_case__ : Tuple = None snake_case__ : Tuple = None snake_case__ : List[str] = None self.builder.download_and_prepare( download_config=snake_case_ , download_mode=snake_case_ , verification_mode=snake_case_ , base_path=snake_case_ , num_proc=self.num_proc , ) snake_case__ : Dict = self.builder.as_dataset( split=self.split , verification_mode=snake_case_ , in_memory=self.keep_in_memory ) return dataset class UpperCAmelCase_ : """simple docstring""" def __init__( self : List[str] , snake_case_ : Dataset , snake_case_ : Union[PathLike, BinaryIO] , snake_case_ : Optional[int] = None , snake_case_ : Optional[int] = None , **snake_case_ : List[Any] , ): if num_proc is not None and num_proc <= 0: raise ValueError(f"num_proc {num_proc} must be an integer > 0." ) snake_case__ : Optional[int] = dataset snake_case__ : int = path_or_buf snake_case__ : Union[str, Any] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE snake_case__ : int = num_proc snake_case__ : Optional[int] = """utf-8""" snake_case__ : Any = to_json_kwargs def lowerCamelCase ( self : List[Any] ): snake_case__ : List[Any] = self.to_json_kwargs.pop("""path_or_buf""" , snake_case_ ) snake_case__ : List[Any] = self.to_json_kwargs.pop("""orient""" , """records""" ) snake_case__ : Any = self.to_json_kwargs.pop("""lines""" , True if orient == """records""" else False ) snake_case__ : Union[str, Any] = self.to_json_kwargs.pop("""index""" , False if orient in ["""split""", """table"""] else True ) snake_case__ : Tuple = self.to_json_kwargs.pop("""compression""" , snake_case_ ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(f"`datasets` currently does not support {compression} compression" ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , """wb""" , compression=snake_case_ ) as buffer: snake_case__ : List[Any] = self._write(file_obj=snake_case_ , orient=snake_case_ , lines=snake_case_ , index=snake_case_ , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( f"The compression parameter is not supported when writing to a buffer, but compression={compression}" """ was passed. Please provide a local path instead.""" ) snake_case__ : str = self._write( file_obj=self.path_or_buf , orient=snake_case_ , lines=snake_case_ , index=snake_case_ , **self.to_json_kwargs ) return written def lowerCamelCase ( self : str , snake_case_ : Dict ): snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ : Any = args snake_case__ : Optional[int] = query_table( table=self.dataset.data , key=slice(snake_case_ , offset + self.batch_size ) , indices=self.dataset._indices , ) snake_case__ : int = batch.to_pandas().to_json( path_or_buf=snake_case_ , orient=snake_case_ , lines=snake_case_ , index=snake_case_ , **snake_case_ ) if not json_str.endswith("""\n""" ): json_str += "\n" return json_str.encode(self.encoding ) def lowerCamelCase ( self : Tuple , snake_case_ : BinaryIO , snake_case_ : List[Any] , snake_case_ : str , snake_case_ : Dict , **snake_case_ : Dict , ): snake_case__ : Tuple = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): snake_case__ : List[Any] = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(snake_case_ ) else: snake_case__ , snake_case__ : Optional[int] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , snake_case_ , snake_case_ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): written += file_obj.write(snake_case_ ) return written
35
'''simple docstring''' import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase_ ( _a , unittest.TestCase ): """simple docstring""" lowercase = GPTSanJapaneseTokenizer lowercase = False lowercase = {"do_clean_text": False, "add_prefix_space": False} def lowerCamelCase ( self : str ): super().setUp() # fmt: off snake_case__ : Optional[Any] = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on snake_case__ : int = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 snake_case__ : List[Any] = {"""unk_token""": """<unk>"""} snake_case__ : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) snake_case__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(snake_case_ ) ) def lowerCamelCase ( self : Any , **snake_case_ : Union[str, Any] ): kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def lowerCamelCase ( self : Any , snake_case_ : str ): snake_case__ : Union[str, Any] = """こんにちは、世界。 \nこんばんは、㔺界。😀""" snake_case__ : List[str] = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def lowerCamelCase ( self : Any , snake_case_ : Dict ): snake_case__ , snake_case__ : int = self.get_input_output_texts(snake_case_ ) snake_case__ : int = tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) snake_case__ : List[str] = tokenizer.decode(snake_case_ , clean_up_tokenization_spaces=snake_case_ ) return text, ids def lowerCamelCase ( self : Optional[Any] ): pass # TODO add if relevant def lowerCamelCase ( self : Union[str, Any] ): pass # TODO add if relevant def lowerCamelCase ( self : List[str] ): pass # TODO add if relevant def lowerCamelCase ( self : Dict ): snake_case__ : Optional[Any] = self.get_tokenizer() # Testing tokenization snake_case__ : int = """こんにちは、世界。 こんばんは、㔺界。""" snake_case__ : Optional[int] = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] snake_case__ : Dict = tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) # Testing conversion to ids without special tokens snake_case__ : Union[str, Any] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] snake_case__ : List[Any] = tokenizer.convert_tokens_to_ids(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) # Testing conversion to ids with special tokens snake_case__ : Union[str, Any] = tokens + [tokenizer.unk_token] snake_case__ : Dict = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] snake_case__ : Any = tokenizer.convert_tokens_to_ids(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) def lowerCamelCase ( self : Optional[Any] ): snake_case__ : Union[str, Any] = self.get_tokenizer() # Testing tokenization snake_case__ : Union[str, Any] = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" snake_case__ : Optional[int] = """こんにちは、、、、世界。こんばんは、、、、世界。""" snake_case__ : Any = tokenizer.encode(snake_case_ ) snake_case__ : int = tokenizer.decode(snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) @slow def lowerCamelCase ( self : Union[str, Any] ): snake_case__ : Optional[Any] = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization snake_case__ : Tuple = """こんにちは、世界。""" snake_case__ : Optional[Any] = """こんばんは、㔺界。😀""" snake_case__ : List[str] = """こんにちは、世界。こんばんは、世界。😀""" snake_case__ : Dict = tokenizer.encode(prefix_text + input_text ) snake_case__ : Dict = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) snake_case__ : int = tokenizer.encode(snake_case_ , prefix_text=snake_case_ ) snake_case__ : Optional[Any] = tokenizer.decode(snake_case_ ) snake_case__ : Union[str, Any] = tokenizer.decode(snake_case_ ) snake_case__ : str = tokenizer.decode(snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) @slow def lowerCamelCase ( self : Union[str, Any] ): snake_case__ : Optional[Any] = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization snake_case__ : Dict = """こんにちは、世界。""" snake_case__ : Optional[int] = """こんばんは、㔺界。😀""" snake_case__ : Any = len(tokenizer.encode(snake_case_ ) ) - 2 snake_case__ : Optional[int] = len(tokenizer.encode(snake_case_ ) ) - 2 snake_case__ : List[str] = [1] + [0] * (len_prefix + len_text + 1) snake_case__ : Optional[int] = [1] * (len_prefix + len_text + 1) + [0] snake_case__ : int = [1] + [1] * (len_prefix) + [0] * (len_text + 1) snake_case__ : Any = tokenizer(prefix_text + input_text ).token_type_ids snake_case__ : str = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids snake_case__ : Optional[Any] = tokenizer(snake_case_ , prefix_text=snake_case_ ).token_type_ids self.assertListEqual(snake_case_ , snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def lowerCamelCase ( self : Optional[int] ): snake_case__ : Optional[Any] = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) snake_case__ : Union[str, Any] = tokenizer.encode("""あンいワ""" ) snake_case__ : int = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) snake_case__ : Dict = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(snake_case_ ) , tokenizer.decode(snake_case_ ) ) self.assertEqual(tokenizer.decode(snake_case_ ) , tokenizer.decode(snake_case_ ) ) self.assertNotEqual(snake_case_ , snake_case_ ) self.assertNotEqual(snake_case_ , snake_case_ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def lowerCamelCase ( self : Any ): snake_case__ : Optional[int] = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) snake_case__ : int = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] snake_case__ : Optional[Any] = tokenizer(snake_case_ , padding=snake_case_ ) snake_case__ : Tuple = tokenizer.batch_encode_plus(snake_case_ , padding=snake_case_ ) # fmt: off snake_case__ : Optional[Any] = [[35_993, 8_640, 25_948, 35_998, 30_647, 35_675, 35_999, 35_999], [35_993, 10_382, 9_868, 35_998, 30_646, 9_459, 30_646, 35_675]] snake_case__ : Optional[Any] = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] snake_case__ : Optional[Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , snake_case_ ) self.assertListEqual(x_token.token_type_ids , snake_case_ ) self.assertListEqual(x_token.attention_mask , snake_case_ ) self.assertListEqual(x_token_a.input_ids , snake_case_ ) self.assertListEqual(x_token_a.token_type_ids , snake_case_ ) self.assertListEqual(x_token_a.attention_mask , snake_case_ ) def lowerCamelCase ( self : Any ): # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def lowerCamelCase ( self : List[str] ): # tokenizer has no padding token pass
35
1
"""simple docstring""" from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
150
"""simple docstring""" from __future__ import annotations class __UpperCAmelCase: """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' lowercase__ : List[Any]= data lowercase__ : Node | None= None lowercase__ : Node | None= None def lowercase__(A ) ->None: # In Order traversal of the tree """simple docstring""" if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowercase__(A ) ->int: """simple docstring""" return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowercase__(A ) ->bool: """simple docstring""" if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowercase__() ->None: # Main function for testing. """simple docstring""" lowercase__ : Tuple= Node(1 ) lowercase__ : Optional[int]= Node(2 ) lowercase__ : List[str]= Node(3 ) lowercase__ : Tuple= Node(4 ) lowercase__ : Optional[int]= Node(5 ) lowercase__ : Any= Node(6 ) lowercase__ : Optional[Any]= Node(7 ) lowercase__ : Optional[int]= Node(8 ) lowercase__ : List[str]= Node(9 ) print(is_full_binary_tree(A ) ) print(depth_of_tree(A ) ) print("Tree is: " ) display(A ) if __name__ == "__main__": main()
150
1
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES lowercase_ = logging.get_logger(__name__) lowercase_ = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) lowercase_ = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) lowercase_ = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) lowercase_ = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) lowercase_ = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) lowercase_ = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) lowercase_ = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) lowercase_ = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) lowercase_ = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) lowercase_ = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) lowercase_ = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) lowercase_ = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) lowercase_ = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) lowercase_ = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) lowercase_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_MAPPING lowercase_ = auto_class_update(FlaxAutoModel) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_PRETRAINING_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_MASKED_LM_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING lowercase_ = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class __lowerCAmelCase ( _BaseAutoModelClass ): _a = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING lowercase_ = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
205
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( _lowerCamelCase : int ) -> list[int]: lowerCamelCase_ = [True] * limit lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): lowerCamelCase_ = i * 2 while index < limit: lowerCamelCase_ = False lowerCamelCase_ = index + i lowerCamelCase_ = [2] for i in range(3 , _lowerCamelCase , 2 ): if is_prime[i]: primes.append(_lowerCamelCase ) return primes def lowerCamelCase__ ( _lowerCamelCase : int = 1000000 ) -> int: lowerCamelCase_ = prime_sieve(_lowerCamelCase ) lowerCamelCase_ = 0 lowerCamelCase_ = 0 for i in range(len(_lowerCamelCase ) ): for j in range(i + length , len(_lowerCamelCase ) ): lowerCamelCase_ = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: lowerCamelCase_ = j - i lowerCamelCase_ = sol return largest if __name__ == "__main__": print(F'''{solution() = }''')
183
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __lowerCAmelCase ( _a ): lowerCamelCase_ : Union[List[np.ndarray], torch.FloatTensor] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
279
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { '''microsoft/biogpt''': '''https://huggingface.co/microsoft/biogpt/resolve/main/config.json''', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __lowerCAmelCase ( _a ): lowerCamelCase_ : Any = '''biogpt''' def __init__(self , __magic_name__=4_2384 , __magic_name__=1024 , __magic_name__=24 , __magic_name__=16 , __magic_name__=4096 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=1024 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=True , __magic_name__=True , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , **__magic_name__ , ) -> List[str]: '''simple docstring''' snake_case_ : List[str] = vocab_size snake_case_ : Dict = max_position_embeddings snake_case_ : Optional[int] = hidden_size snake_case_ : List[Any] = num_hidden_layers snake_case_ : List[str] = num_attention_heads snake_case_ : int = intermediate_size snake_case_ : List[Any] = hidden_act snake_case_ : List[Any] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : Optional[int] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = scale_embedding snake_case_ : Optional[Any] = use_cache snake_case_ : Optional[Any] = layerdrop snake_case_ : Optional[Any] = activation_dropout super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ )
279
1
"""simple docstring""" _a = '\n# Installazione di Transformers\n! pip install transformers datasets\n# Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e\n# rimuovi la modalità commento al comando seguente.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _a = [{'type': 'code', 'content': INSTALL_CONTENT}] _a = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
61
"""simple docstring""" _a : List[str] = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
"""simple docstring""" from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class a ( lowerCAmelCase_ ): _snake_case : Tuple = '' _snake_case : List[str] = 'hf-legacy' # "hf://"" is reserved for hffs def __init__( self : Union[str, Any] , __lowerCAmelCase : Optional[DatasetInfo] = None , __lowerCAmelCase : Optional[str] = None , **__lowerCAmelCase : Union[str, Any] , ): super().__init__(self , **__lowerCAmelCase ) _UpperCAmelCase = repo_info _UpperCAmelCase = token _UpperCAmelCase = None def lowerCAmelCase_ ( self : int ): if self.dir_cache is None: _UpperCAmelCase = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes _UpperCAmelCase = { """name""": hf_file.rfilename, """size""": None, """type""": """file""", } self.dir_cache.update( { str(__lowerCAmelCase ): {"""name""": str(__lowerCAmelCase ), """size""": None, """type""": """directory"""} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def lowerCAmelCase_ ( self : Dict , __lowerCAmelCase : str , __lowerCAmelCase : str = "rb" , **__lowerCAmelCase : List[Any] , ): if not isinstance(self.repo_info , __lowerCAmelCase ): raise NotImplementedError(f'''Open is only implemented for dataset repositories, but got {self.repo_info}''' ) _UpperCAmelCase = hf_hub_url(self.repo_info.id , __lowerCAmelCase , revision=self.repo_info.sha ) return fsspec.open( __lowerCAmelCase , mode=__lowerCAmelCase , headers=get_authentication_headers_for_url(__lowerCAmelCase , use_auth_token=self.token ) , client_kwargs={"""trust_env""": True} , ).open() def lowerCAmelCase_ ( self : List[Any] , __lowerCAmelCase : Any , **__lowerCAmelCase : Dict ): self._get_dirs() _UpperCAmelCase = self._strip_protocol(__lowerCAmelCase ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(__lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[Any]=False , **__lowerCAmelCase : Tuple ): self._get_dirs() _UpperCAmelCase = PurePosixPath(path.strip("""/""" ) ) _UpperCAmelCase = {} for p, f in self.dir_cache.items(): _UpperCAmelCase = PurePosixPath(p.strip("""/""" ) ) _UpperCAmelCase = p.parent if root == path: _UpperCAmelCase = f _UpperCAmelCase = list(paths.values() ) if detail: return out else: return sorted(f["""name"""] for f in out )
30
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) UpperCAmelCase__ = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
30
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A__ = logging.get_logger(__name__) A__ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCamelCase__ ): __lowerCamelCase = '''poolformer''' def __init__( self , _snake_case=3 , _snake_case=16 , _snake_case=16 , _snake_case=3 , _snake_case=4.0 , _snake_case=[2, 2, 6, 2] , _snake_case=[64, 128, 320, 512] , _snake_case=[7, 3, 3, 3] , _snake_case=[4, 2, 2, 2] , _snake_case=[2, 1, 1, 1] , _snake_case=4 , _snake_case=0.0 , _snake_case="gelu" , _snake_case=True , _snake_case=1e-5 , _snake_case=0.02 , **_snake_case , ): """simple docstring""" _lowerCAmelCase = num_channels _lowerCAmelCase = patch_size _lowerCAmelCase = stride _lowerCAmelCase = padding _lowerCAmelCase = pool_size _lowerCAmelCase = hidden_sizes _lowerCAmelCase = mlp_ratio _lowerCAmelCase = depths _lowerCAmelCase = patch_sizes _lowerCAmelCase = strides _lowerCAmelCase = num_encoder_blocks _lowerCAmelCase = drop_path_rate _lowerCAmelCase = hidden_act _lowerCAmelCase = use_layer_scale _lowerCAmelCase = layer_scale_init_value _lowerCAmelCase = initializer_range super().__init__(**_snake_case ) class __lowerCAmelCase ( lowerCamelCase__ ): __lowerCamelCase = version.parse('''1.11''' ) @property def snake_case ( self ): """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case ( self ): """simple docstring""" return 2e-3
82
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class __lowerCAmelCase ( unittest.TestCase ): def snake_case ( self ): """simple docstring""" _lowerCAmelCase = { """task_specific_params""": { """summarization""": {"""length_penalty""": 1.0, """max_length""": 128, """min_length""": 12, """num_beams""": 4}, """summarization_cnn""": {"""length_penalty""": 2.0, """max_length""": 142, """min_length""": 56, """num_beams""": 4}, """summarization_xsum""": {"""length_penalty""": 1.0, """max_length""": 62, """min_length""": 11, """num_beams""": 6}, } } _lowerCAmelCase = { """task_specific_params.summarization.length_penalty""": 1.0, """task_specific_params.summarization.max_length""": 128, """task_specific_params.summarization.min_length""": 12, """task_specific_params.summarization.num_beams""": 4, """task_specific_params.summarization_cnn.length_penalty""": 2.0, """task_specific_params.summarization_cnn.max_length""": 142, """task_specific_params.summarization_cnn.min_length""": 56, """task_specific_params.summarization_cnn.num_beams""": 4, """task_specific_params.summarization_xsum.length_penalty""": 1.0, """task_specific_params.summarization_xsum.max_length""": 62, """task_specific_params.summarization_xsum.min_length""": 11, """task_specific_params.summarization_xsum.num_beams""": 6, } self.assertEqual(flatten_dict(_snake_case ) , _snake_case ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(_snake_case ) , x.transpose() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , transpose(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , transpose(_snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , transpose(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , transpose(_snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , np.asarray(transpose(_snake_case ) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(_snake_case , axes=(1, 2, 0) ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , np.reshape(_snake_case , (4, 3) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , np.reshape(_snake_case , (12, 5) ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , reshape(_snake_case , (4, 3) ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , reshape(_snake_case , (12, 5) ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , reshape(_snake_case , (4, 3) ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , reshape(_snake_case , (12, 5) ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , np.asarray(reshape(_snake_case , (4, 3) ) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , np.asarray(reshape(_snake_case , (12, 5) ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(_snake_case ) , np.squeeze(_snake_case ) ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , np.squeeze(_snake_case , axis=2 ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , squeeze(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , squeeze(_snake_case , axis=2 ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , squeeze(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , squeeze(_snake_case , axis=2 ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , np.asarray(squeeze(_snake_case ) ) ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , np.asarray(squeeze(_snake_case , axis=2 ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , np.expand_dims(_snake_case , axis=1 ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , expand_dims(_snake_case , axis=1 ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , expand_dims(_snake_case , axis=1 ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , np.asarray(expand_dims(_snake_case , axis=1 ) ) ) )
82
1
'''simple docstring''' from typing import List from .keymap import KEYMAP, get_character def __UpperCamelCase ( UpperCAmelCase ): def decorator(UpperCAmelCase ): lowercase__ : Dict = getattr(UpperCAmelCase , '''handle_key''' , [] ) handle += [key] setattr(UpperCAmelCase , '''handle_key''' , UpperCAmelCase ) return func return decorator def __UpperCamelCase ( *UpperCAmelCase ): def decorator(UpperCAmelCase ): lowercase__ : Union[str, Any] = getattr(UpperCAmelCase , '''handle_key''' , [] ) handle += keys setattr(UpperCAmelCase , '''handle_key''' , UpperCAmelCase ) return func return decorator class UpperCAmelCase ( a__ ): '''simple docstring''' def __new__( cls , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Any: lowercase__ : List[Any] = super().__new__(cls , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) if not hasattr(__lowerCAmelCase , '''key_handler''' ): setattr(__lowerCAmelCase , '''key_handler''' , {} ) setattr(__lowerCAmelCase , '''handle_input''' , KeyHandler.handle_input ) for value in attrs.values(): lowercase__ : Optional[Any] = getattr(__lowerCAmelCase , '''handle_key''' , [] ) for key in handled_keys: lowercase__ : Tuple = value return new_cls @staticmethod def _lowerCAmelCase( cls ) -> Tuple: lowercase__ : Dict = get_character() if char != KEYMAP["undefined"]: lowercase__ : Tuple = ord(__lowerCAmelCase ) lowercase__ : str = cls.key_handler.get(__lowerCAmelCase ) if handler: lowercase__ : str = char return handler(cls ) else: return None def __UpperCamelCase ( cls ): return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
352
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax __a: Optional[Any] = logging.get_logger(__name__) @add_end_docstrings(a__ ) class UpperCAmelCase ( a__ ): '''simple docstring''' def __init__( self , **__lowerCAmelCase ) -> int: super().__init__(**__lowerCAmelCase ) requires_backends(self , '''vision''' ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self , __lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return super().__call__(__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCAmelCase( self , **__lowerCAmelCase ) -> Optional[Any]: lowercase__ : str = {} if "candidate_labels" in kwargs: lowercase__ : str = kwargs['''candidate_labels'''] if "hypothesis_template" in kwargs: lowercase__ : str = kwargs['''hypothesis_template'''] return preprocess_params, {}, {} def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase="This is a photo of {}." ) -> Any: lowercase__ : Union[str, Any] = load_image(__lowerCAmelCase ) lowercase__ : List[str] = self.image_processor(images=[image] , return_tensors=self.framework ) lowercase__ : Union[str, Any] = candidate_labels lowercase__ : int = [hypothesis_template.format(__lowerCAmelCase ) for x in candidate_labels] lowercase__ : Any = self.tokenizer(__lowerCAmelCase , return_tensors=self.framework , padding=__lowerCAmelCase ) lowercase__ : Any = [text_inputs] return inputs def _lowerCAmelCase( self , __lowerCAmelCase ) -> Optional[Any]: lowercase__ : Any = model_inputs.pop('''candidate_labels''' ) lowercase__ : int = model_inputs.pop('''text_inputs''' ) if isinstance(text_inputs[0] , __lowerCAmelCase ): lowercase__ : Union[str, Any] = text_inputs[0] else: # Batching case. lowercase__ : Optional[Any] = text_inputs[0][0] lowercase__ : Any = self.model(**__lowerCAmelCase , **__lowerCAmelCase ) lowercase__ : Any = { '''candidate_labels''': candidate_labels, '''logits''': outputs.logits_per_image, } return model_outputs def _lowerCAmelCase( self , __lowerCAmelCase ) -> List[str]: lowercase__ : Union[str, Any] = model_outputs.pop('''candidate_labels''' ) lowercase__ : Optional[int] = model_outputs['''logits'''][0] if self.framework == "pt": lowercase__ : Optional[int] = logits.softmax(dim=-1 ).squeeze(-1 ) lowercase__ : Any = probs.tolist() if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): lowercase__ : Dict = [scores] elif self.framework == "tf": lowercase__ : List[Any] = stable_softmax(__lowerCAmelCase , axis=-1 ) lowercase__ : str = probs.numpy().tolist() else: raise ValueError(F"""Unsupported framework: {self.framework}""" ) lowercase__ : Optional[int] = [ {'''score''': score, '''label''': candidate_label} for score, candidate_label in sorted(zip(__lowerCAmelCase , __lowerCAmelCase ) , key=lambda __lowerCAmelCase : -x[0] ) ] return result
214
0
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = "xlm-prophetnet" lowerCAmelCase : Union[str, Any] = ["past_key_values"] lowerCAmelCase : Union[str, Any] = { "num_attention_heads": "num_encoder_attention_heads", } def __init__( self : Dict , lowerCamelCase__ : Optional[float] = 0.1 , lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" , lowerCamelCase__ : Optional[int] = 3_05_22 , lowerCamelCase__ : Optional[int] = 10_24 , lowerCamelCase__ : Optional[int] = 40_96 , lowerCamelCase__ : Optional[int] = 12 , lowerCamelCase__ : Optional[int] = 16 , lowerCamelCase__ : Optional[int] = 40_96 , lowerCamelCase__ : Optional[int] = 12 , lowerCamelCase__ : Optional[int] = 16 , lowerCamelCase__ : Optional[float] = 0.1 , lowerCamelCase__ : Optional[float] = 0.1 , lowerCamelCase__ : Optional[int] = 5_12 , lowerCamelCase__ : Optional[float] = 0.0_2 , lowerCamelCase__ : Optional[bool] = True , lowerCamelCase__ : Optional[bool] = True , lowerCamelCase__ : Optional[int] = 0 , lowerCamelCase__ : Optional[int] = 2 , lowerCamelCase__ : Optional[int] = 32 , lowerCamelCase__ : Optional[int] = 1_28 , lowerCamelCase__ : Optional[bool] = False , lowerCamelCase__ : Optional[float] = 0.0 , lowerCamelCase__ : Optional[bool] = True , lowerCamelCase__ : Optional[int] = 0 , lowerCamelCase__ : Optional[int] = 1 , lowerCamelCase__ : Optional[int] = 2 , **lowerCamelCase__ : Optional[int] , ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Dict = vocab_size _UpperCAmelCase : Union[str, Any] = hidden_size _UpperCAmelCase : Optional[int] = encoder_ffn_dim _UpperCAmelCase : Tuple = num_encoder_layers _UpperCAmelCase : str = num_encoder_attention_heads _UpperCAmelCase : Dict = decoder_ffn_dim _UpperCAmelCase : List[Any] = num_decoder_layers _UpperCAmelCase : str = num_decoder_attention_heads _UpperCAmelCase : Any = max_position_embeddings _UpperCAmelCase : Union[str, Any] = init_std # Normal(0, this parameter) _UpperCAmelCase : Tuple = activation_function # parameters for xlmprophetnet _UpperCAmelCase : List[Any] = ngram _UpperCAmelCase : Optional[int] = num_buckets _UpperCAmelCase : List[str] = relative_max_distance _UpperCAmelCase : List[Any] = disable_ngram_loss _UpperCAmelCase : Tuple = eps # 3 Types of Dropout _UpperCAmelCase : Any = attention_dropout _UpperCAmelCase : Tuple = activation_dropout _UpperCAmelCase : Optional[int] = dropout _UpperCAmelCase : Optional[Any] = use_cache super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , is_encoder_decoder=lowerCamelCase__ , add_cross_attention=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , ) @property def lowerCAmelCase__ ( self : Any ) ->int: '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict ) ->str: '''simple docstring''' raise NotImplementedError( "This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and" " `num_decoder_layers`." )
234
'''simple docstring''' from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo lowerCamelCase__ = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey\n and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin\n Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto\n Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and\n Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes\n and Jeffrey Dean},\n year={2016},\n eprint={1609.08144},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n' lowerCamelCase__ = '\\nThe BLEU score has some undesirable properties when used for single\nsentences, as it was designed to be a corpus measure. We therefore\nuse a slightly different score for our RL experiments which we call\nthe \'GLEU score\'. For the GLEU score, we record all sub-sequences of\n1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then\ncompute a recall, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the target (ground truth) sequence,\nand a precision, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the generated output sequence. Then\nGLEU score is simply the minimum of recall and precision. This GLEU\nscore\'s range is always between 0 (no matches) and 1 (all match) and\nit is symmetrical when switching output and target. According to\nour experiments, GLEU score correlates quite well with the BLEU\nmetric on a corpus level but does not have its drawbacks for our per\nsentence reward objective.\n' lowerCamelCase__ = '\\nComputes corpus-level Google BLEU (GLEU) score of translated segments against one or more references.\nInstead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching\ntokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values.\n\nArgs:\n predictions (list of str): list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references (list of list of str): list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n min_len (int): The minimum order of n-gram this function should extract. Defaults to 1.\n max_len (int): The maximum order of n-gram this function should extract. Defaults to 4.\n\nReturns:\n \'google_bleu\': google_bleu score\n\nExamples:\n Example 1:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.44\n\n Example 2:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.61\n\n Example 3:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2)\n >>> print(round(results["google_bleu"], 2))\n 0.53\n\n Example 4:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6)\n >>> print(round(results["google_bleu"], 2))\n 0.4\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : int ) ->MetricInfo: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ), "references": datasets.Sequence( datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ) , id="references" ), } ) , ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[List[List[str]]] , lowerCamelCase__ : List[List[str]] , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 4 , ) ->Dict[str, float]: '''simple docstring''' return { "google_bleu": gleu_score.corpus_gleu( list_of_references=lowerCamelCase__ , hypotheses=lowerCamelCase__ , min_len=lowerCamelCase__ , max_len=lowerCamelCase__ ) }
234
1
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def UpperCamelCase( UpperCAmelCase_ ): UpperCAmelCase : List[str] = SwinaSRConfig() if "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: UpperCAmelCase : str = 4 elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: UpperCAmelCase : Tuple = 4 UpperCAmelCase : Dict = 48 UpperCAmelCase : Union[str, Any] = 'pixelshuffle_aux' elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: UpperCAmelCase : Tuple = [6, 6, 6, 6] UpperCAmelCase : Optional[Any] = 60 UpperCAmelCase : Union[str, Any] = [6, 6, 6, 6] UpperCAmelCase : Tuple = 'pixelshuffledirect' elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: UpperCAmelCase : Union[str, Any] = 4 UpperCAmelCase : Optional[int] = 'nearest+conv' elif "Swin2SR_Jpeg_dynamic" in checkpoint_url: UpperCAmelCase : Tuple = 1 UpperCAmelCase : Optional[int] = 1 UpperCAmelCase : Union[str, Any] = 1_26 UpperCAmelCase : str = 7 UpperCAmelCase : List[str] = 255.0 UpperCAmelCase : Optional[int] = '' return config def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): if "patch_embed.proj" in name and "layers" not in name: UpperCAmelCase : List[str] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: UpperCAmelCase : List[str] = name.replace('patch_embed.norm' , 'embeddings.patch_embeddings.layernorm' ) if "layers" in name: UpperCAmelCase : Any = name.replace('layers' , 'encoder.stages' ) if "residual_group.blocks" in name: UpperCAmelCase : Union[str, Any] = name.replace('residual_group.blocks' , 'layers' ) if "attn.proj" in name: UpperCAmelCase : Optional[Any] = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: UpperCAmelCase : List[str] = name.replace('attn' , 'attention.self' ) if "norm1" in name: UpperCAmelCase : List[str] = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: UpperCAmelCase : List[Any] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: UpperCAmelCase : Tuple = name.replace('mlp.fc2' , 'output.dense' ) if "q_bias" in name: UpperCAmelCase : Dict = name.replace('q_bias' , 'query.bias' ) if "k_bias" in name: UpperCAmelCase : str = name.replace('k_bias' , 'key.bias' ) if "v_bias" in name: UpperCAmelCase : List[str] = name.replace('v_bias' , 'value.bias' ) if "cpb_mlp" in name: UpperCAmelCase : Union[str, Any] = name.replace('cpb_mlp' , 'continuous_position_bias_mlp' ) if "patch_embed.proj" in name: UpperCAmelCase : Optional[Any] = name.replace('patch_embed.proj' , 'patch_embed.projection' ) if name == "norm.weight": UpperCAmelCase : List[str] = 'layernorm.weight' if name == "norm.bias": UpperCAmelCase : int = 'layernorm.bias' if "conv_first" in name: UpperCAmelCase : Union[str, Any] = name.replace('conv_first' , 'first_convolution' ) if ( "upsample" in name or "conv_before_upsample" in name or "conv_bicubic" in name or "conv_up" in name or "conv_hr" in name or "conv_last" in name or "aux" in name ): # heads if "conv_last" in name: UpperCAmelCase : Dict = name.replace('conv_last' , 'final_convolution' ) if config.upsampler in ["pixelshuffle", "pixelshuffle_aux", "nearest+conv"]: if "conv_before_upsample.0" in name: UpperCAmelCase : int = name.replace('conv_before_upsample.0' , 'conv_before_upsample' ) if "upsample.0" in name: UpperCAmelCase : Optional[Any] = name.replace('upsample.0' , 'upsample.convolution_0' ) if "upsample.2" in name: UpperCAmelCase : List[str] = name.replace('upsample.2' , 'upsample.convolution_1' ) UpperCAmelCase : List[Any] = 'upsample.' + name elif config.upsampler == "pixelshuffledirect": UpperCAmelCase : Tuple = name.replace('upsample.0.weight' , 'upsample.conv.weight' ) UpperCAmelCase : Dict = name.replace('upsample.0.bias' , 'upsample.conv.bias' ) else: pass else: UpperCAmelCase : List[Any] = 'swin2sr.' + name return name def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): for key in orig_state_dict.copy().keys(): UpperCAmelCase : int = orig_state_dict.pop(UpperCAmelCase_ ) if "qkv" in key: UpperCAmelCase : Optional[int] = key.split('.' ) UpperCAmelCase : Tuple = int(key_split[1] ) UpperCAmelCase : Union[str, Any] = int(key_split[4] ) UpperCAmelCase : Any = config.embed_dim if "weight" in key: UpperCAmelCase : Tuple = val[:dim, :] UpperCAmelCase : List[str] = val[dim : dim * 2, :] UpperCAmelCase : List[Any] = val[-dim:, :] else: UpperCAmelCase : str = val[:dim] UpperCAmelCase : Any = val[dim : dim * 2] UpperCAmelCase : Optional[Any] = val[-dim:] pass else: UpperCAmelCase : Tuple = val return orig_state_dict def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): UpperCAmelCase : int = get_config(UpperCAmelCase_ ) UpperCAmelCase : List[Any] = SwinaSRForImageSuperResolution(UpperCAmelCase_ ) model.eval() UpperCAmelCase : Union[str, Any] = torch.hub.load_state_dict_from_url(UpperCAmelCase_ , map_location='cpu' ) UpperCAmelCase : Tuple = convert_state_dict(UpperCAmelCase_ , UpperCAmelCase_ ) UpperCAmelCase , UpperCAmelCase : List[Any] = model.load_state_dict(UpperCAmelCase_ , strict=UpperCAmelCase_ ) if len(UpperCAmelCase_ ) > 0: raise ValueError('Missing keys when converting: {}'.format(UpperCAmelCase_ ) ) for key in unexpected_keys: if not ("relative_position_index" in key or "relative_coords_table" in key or "self_mask" in key): raise ValueError(F"""Unexpected key {key} in state_dict""" ) # verify values UpperCAmelCase : Optional[Any] = 'https://github.com/mv-lab/swin2sr/blob/main/testsets/real-inputs/shanghai.jpg?raw=true' UpperCAmelCase : Any = Image.open(requests.get(UpperCAmelCase_ , stream=UpperCAmelCase_ ).raw ).convert('RGB' ) UpperCAmelCase : Dict = SwinaSRImageProcessor() # pixel_values = processor(image, return_tensors="pt").pixel_values UpperCAmelCase : Optional[Any] = 1_26 if 'Jpeg' in checkpoint_url else 2_56 UpperCAmelCase : List[str] = Compose( [ Resize((image_size, image_size) ), ToTensor(), Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) UpperCAmelCase : Union[str, Any] = transforms(UpperCAmelCase_ ).unsqueeze(0 ) if config.num_channels == 1: UpperCAmelCase : Dict = pixel_values[:, 0, :, :].unsqueeze(1 ) UpperCAmelCase : List[Any] = model(UpperCAmelCase_ ) # assert values if "Swin2SR_ClassicalSR_X2_64" in checkpoint_url: UpperCAmelCase : Any = torch.Size([1, 3, 5_12, 5_12] ) UpperCAmelCase : Optional[Any] = torch.tensor( [[-0.7087, -0.7138, -0.6721], [-0.8340, -0.8095, -0.7298], [-0.9149, -0.8414, -0.7940]] ) elif "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: UpperCAmelCase : str = torch.Size([1, 3, 10_24, 10_24] ) UpperCAmelCase : int = torch.tensor( [[-0.7775, -0.8105, -0.8933], [-0.7764, -0.8356, -0.9225], [-0.7976, -0.8686, -0.9579]] ) elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: # TODO values didn't match exactly here UpperCAmelCase : List[Any] = torch.Size([1, 3, 10_24, 10_24] ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[-0.8035, -0.7504, -0.7491], [-0.8538, -0.8124, -0.7782], [-0.8804, -0.8651, -0.8493]] ) elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: UpperCAmelCase : Any = torch.Size([1, 3, 5_12, 5_12] ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[-0.7669, -0.8662, -0.8767], [-0.8810, -0.9962, -0.9820], [-0.9340, -1.0322, -1.1149]] ) elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: UpperCAmelCase : Optional[Any] = torch.Size([1, 3, 10_24, 10_24] ) UpperCAmelCase : str = torch.tensor( [[-0.5238, -0.5557, -0.6321], [-0.6016, -0.5903, -0.6391], [-0.6244, -0.6334, -0.6889]] ) assert ( outputs.reconstruction.shape == expected_shape ), F"""Shape of reconstruction should be {expected_shape}, but is {outputs.reconstruction.shape}""" assert torch.allclose(outputs.reconstruction[0, 0, :3, :3] , UpperCAmelCase_ , atol=1E-3 ) print('Looks ok!' ) UpperCAmelCase : List[Any] = { 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth': ( 'swin2SR-classical-sr-x2-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X4_64.pth': ( 'swin2SR-classical-sr-x4-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_CompressedSR_X4_48.pth': ( 'swin2SR-compressed-sr-x4-48' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pth': ( 'swin2SR-lightweight-x2-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR.pth': ( 'swin2SR-realworld-sr-x4-64-bsrgan-psnr' ), } UpperCAmelCase : Optional[int] = url_to_name[checkpoint_url] if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(UpperCAmelCase_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(UpperCAmelCase_ ) if push_to_hub: model.push_to_hub(F"""caidas/{model_name}""" ) processor.push_to_hub(F"""caidas/{model_name}""" ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth", type=str, help="URL of the original Swin2SR checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the converted model to the hub.") lowercase__ = parser.parse_args() convert_swinasr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
280
'''simple docstring''' def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): if b == 0: return 1 if (b % 2) == 0: return actual_power(UpperCAmelCase_ , int(b / 2 ) ) * actual_power(UpperCAmelCase_ , int(b / 2 ) ) else: return a * actual_power(UpperCAmelCase_ , int(b / 2 ) ) * actual_power(UpperCAmelCase_ , int(b / 2 ) ) def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): if b < 0: return 1 / actual_power(UpperCAmelCase_ , UpperCAmelCase_ ) return actual_power(UpperCAmelCase_ , UpperCAmelCase_ ) if __name__ == "__main__": print(power(-2, -3))
280
1
'''simple docstring''' import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : List[Any] = {"vocab_file": "spiece.model"} a : Optional[Any] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } a : Union[str, Any] = { "albert-base-v1": 5_12, "albert-large-v1": 5_12, "albert-xlarge-v1": 5_12, "albert-xxlarge-v1": 5_12, "albert-base-v2": 5_12, "albert-large-v2": 5_12, "albert-xlarge-v2": 5_12, "albert-xxlarge-v2": 5_12, } a : Optional[Any] = "▁" class UpperCamelCase__ ( lowercase__ ): """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , snake_case , snake_case=True , snake_case=True , snake_case=False , snake_case="[CLS]" , snake_case="[SEP]" , snake_case="<unk>" , snake_case="[SEP]" , snake_case="<pad>" , snake_case="[CLS]" , snake_case="[MASK]" , snake_case = None , **snake_case , ): '''simple docstring''' UpperCAmelCase : int = ( AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case , normalized=snake_case ) if isinstance(snake_case , snake_case ) else mask_token ) UpperCAmelCase : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) UpperCAmelCase : Union[str, Any] = do_lower_case UpperCAmelCase : Dict = remove_space UpperCAmelCase : Any = keep_accents UpperCAmelCase : int = vocab_file UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def A_ ( self ): '''simple docstring''' return len(self.sp_model ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' UpperCAmelCase : List[str] = self.__dict__.copy() UpperCAmelCase : Dict = None return state def __setstate__( self , snake_case ): '''simple docstring''' UpperCAmelCase : Tuple = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase : int = {} UpperCAmelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A_ ( self , snake_case ): '''simple docstring''' if self.remove_space: UpperCAmelCase : Optional[int] = " ".join(inputs.strip().split() ) else: UpperCAmelCase : Union[str, Any] = inputs UpperCAmelCase : Optional[int] = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: UpperCAmelCase : Union[str, Any] = unicodedata.normalize("NFKD" , snake_case ) UpperCAmelCase : str = "".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: UpperCAmelCase : int = outputs.lower() return outputs def A_ ( self , snake_case ): '''simple docstring''' UpperCAmelCase : Optional[int] = self.preprocess_text(snake_case ) UpperCAmelCase : Tuple = self.sp_model.encode(snake_case , out_type=snake_case ) UpperCAmelCase : List[Any] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): UpperCAmelCase : str = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: UpperCAmelCase : str = cur_pieces[1:] else: UpperCAmelCase : Union[str, Any] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def A_ ( self , snake_case ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def A_ ( self , snake_case ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def A_ ( self , snake_case ): '''simple docstring''' UpperCAmelCase : int = [] UpperCAmelCase : Dict = "" UpperCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case ) + token UpperCAmelCase : List[str] = True UpperCAmelCase : int = [] else: current_sub_tokens.append(snake_case ) UpperCAmelCase : Union[str, Any] = False out_string += self.sp_model.decode(snake_case ) return out_string.strip() def A_ ( self , snake_case , snake_case = None ): '''simple docstring''' UpperCAmelCase : Dict = [self.sep_token_id] UpperCAmelCase : List[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A_ ( self , snake_case , snake_case = None , snake_case = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return [1] + ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1] return [1] + ([0] * len(snake_case )) + [1] def A_ ( self , snake_case , snake_case = None ): '''simple docstring''' UpperCAmelCase : str = [self.sep_token_id] UpperCAmelCase : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self , snake_case , snake_case = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase : int = os.path.join( snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , "wb" ) as fi: UpperCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
311
'''simple docstring''' def lowercase ( __magic_name__ ): '''simple docstring''' if number > 0: raise ValueError("input must be a negative integer" ) UpperCAmelCase : List[Any] = len(bin(__magic_name__ )[3:] ) UpperCAmelCase : Optional[Any] = bin(abs(__magic_name__ ) - (1 << binary_number_length) )[3:] UpperCAmelCase : Tuple = ( ( "1" + "0" * (binary_number_length - len(__magic_name__ )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
311
1
def __lowerCamelCase ( snake_case__=2_81_23 ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [1] * (limit + 1) for i in range(2 ,int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 ,limit // i + 1 ): sum_divs[k * i] += k + i _SCREAMING_SNAKE_CASE = set() _SCREAMING_SNAKE_CASE = 0 for n in range(1 ,limit + 1 ): if sum_divs[n] > n: abundants.add(snake_case__ ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
125
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[str] = KandinskyVaaInpaintPipeline __snake_case : Union[str, Any] = ["image_embeds", "negative_image_embeds", "image", "mask_image"] __snake_case : Tuple = [ "image_embeds", "negative_image_embeds", "image", "mask_image", ] __snake_case : str = [ "generator", "height", "width", "latents", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] __snake_case : List[str] = False @property def UpperCamelCase ( self: Tuple ): '''simple docstring''' return 32 @property def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' return 32 @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return self.time_input_dim @property def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' return 100 @property def UpperCamelCase ( self: str ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } _SCREAMING_SNAKE_CASE = UNetaDConditionModel(**UpperCAmelCase_ ) return model @property def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = VQModel(**self.dummy_movq_kwargs ) return model def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.dummy_unet _SCREAMING_SNAKE_CASE = self.dummy_movq _SCREAMING_SNAKE_CASE = DDIMScheduler( num_train_timesteps=1_000 , beta_schedule="""linear""" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , steps_offset=1 , prediction_type="""epsilon""" , thresholding=UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str]=0 ): '''simple docstring''' _SCREAMING_SNAKE_CASE = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( UpperCAmelCase_ ) # create init_image _SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = image.cpu().permute(0 , 2 , 3 , 1 )[0] _SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(UpperCAmelCase_ ) ).convert("""RGB""" ).resize((256, 256) ) # create mask _SCREAMING_SNAKE_CASE = np.ones((64, 64) , dtype=np.floataa ) _SCREAMING_SNAKE_CASE = 0 if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = self.pipeline_class(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = pipe.to(UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = pipe(**self.get_dummy_inputs(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = output.images _SCREAMING_SNAKE_CASE = pipe( **self.get_dummy_inputs(UpperCAmelCase_ ) , return_dict=UpperCAmelCase_ , )[0] _SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] _SCREAMING_SNAKE_CASE = image_from_tuple[0, -3:, -3:, -1] print(F'image.shape {image.shape}' ) assert image.shape == (1, 64, 64, 3) _SCREAMING_SNAKE_CASE = np.array( [0.50_77_59_03, 0.49_52_71_95, 0.48_82_45_43, 0.50_19_22_37, 0.48_64_49_06, 0.49_37_38_14, 0.4_78_05_98, 0.47_23_48_27, 0.48_32_78_48] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def UpperCamelCase ( self: int ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) _SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) _SCREAMING_SNAKE_CASE = np.ones((768, 768) , dtype=np.floataa ) _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = """a hat""" _SCREAMING_SNAKE_CASE = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) _SCREAMING_SNAKE_CASE = pipeline.to(UpperCAmelCase_ ) pipeline.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = pipe_prior( UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() _SCREAMING_SNAKE_CASE = pipeline( image=UpperCAmelCase_ , mask_image=UpperCAmelCase_ , image_embeds=UpperCAmelCase_ , negative_image_embeds=UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(UpperCAmelCase_ , UpperCAmelCase_ )
125
1
import warnings from .state import AcceleratorState, GradientState warnings.filterwarnings("ignore", category=UserWarning, module="torch.optim.lr_scheduler") class _lowercase : """simple docstring""" def __init__(self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = True , lowerCamelCase_ = False ): """simple docstring""" a = scheduler a = optimizers if isinstance(lowerCamelCase_ , (list, tuple) ) else [optimizers] a = split_batches a = step_with_optimizer a = GradientState() def UpperCamelCase_ (self , *lowerCamelCase_ , **lowerCamelCase_ ): """simple docstring""" if not self.step_with_optimizer: # No link between scheduler and optimizer -> just step self.scheduler.step(*lowerCamelCase_ , **lowerCamelCase_ ) return # Otherwise, first make sure the optimizer was stepped. if not self.gradient_state.sync_gradients: if self.gradient_state.adjust_scheduler: self.scheduler._step_count += 1 return for opt in self.optimizers: if opt.step_was_skipped: return if self.split_batches: # Split batches -> the training dataloader batch size is not changed so one step per training step self.scheduler.step(*lowerCamelCase_ , **lowerCamelCase_ ) else: # Otherwise the training dataloader batch size was multiplied by `num_processes`, so we need to do # num_processes steps per training step a = AcceleratorState().num_processes for _ in range(lowerCamelCase_ ): # Special case when using OneCycle and `drop_last` was not used if hasattr(self.scheduler , "total_steps" ): if self.scheduler._step_count <= self.scheduler.total_steps: self.scheduler.step(*lowerCamelCase_ , **lowerCamelCase_ ) else: self.scheduler.step(*lowerCamelCase_ , **lowerCamelCase_ ) def UpperCamelCase_ (self ): """simple docstring""" return self.scheduler.get_last_lr() def UpperCamelCase_ (self ): """simple docstring""" return self.scheduler.state_dict() def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" self.scheduler.load_state_dict(lowerCamelCase_ ) def UpperCamelCase_ (self ): """simple docstring""" return self.scheduler.get_lr() def UpperCamelCase_ (self , *lowerCamelCase_ , **lowerCamelCase_ ): """simple docstring""" return self.scheduler.print_lr(*lowerCamelCase_ , **lowerCamelCase_ )
227
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def a ( lowerCamelCase__=None ): '''simple docstring''' A_ : int = argparse.ArgumentParser(add_help=lowerCamelCase__ , allow_abbrev=lowerCamelCase__ ) # The main config parser A_ : int = config_command_parser(lowerCamelCase__ ) # The subparser to add commands to A_ : int = config_parser.add_subparsers(title="""subcommands""" , dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(lowerCamelCase__ , parents=[parent_parser] ) update_command_parser(lowerCamelCase__ , parents=[parent_parser] ) return config_parser def a ( ): '''simple docstring''' A_ : Optional[int] = get_config_parser() A_ : List[str] = config_parser.parse_args() if not hasattr(lowerCamelCase__ , """func""" ): config_parser.print_help() exit(1 ) # Run args.func(lowerCamelCase__ ) if __name__ == "__main__": main()
206
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Union[str, Any] = logging.get_logger(__name__) snake_case : List[Any] = { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _snake_case ( snake_case ): UpperCamelCase__ = 'roformer' def __init__( self , _a=50_000 , _a=None , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=1_536 , _a=2 , _a=0.02 , _a=1e-12 , _a=0 , _a=False , _a=True , **_a , ): super().__init__(pad_token_id=_a , **_a ) __magic_name__ : Tuple = vocab_size __magic_name__ : Dict = hidden_size if embedding_size is None else embedding_size __magic_name__ : int = hidden_size __magic_name__ : int = num_hidden_layers __magic_name__ : Union[str, Any] = num_attention_heads __magic_name__ : Union[str, Any] = hidden_act __magic_name__ : Optional[int] = intermediate_size __magic_name__ : Union[str, Any] = hidden_dropout_prob __magic_name__ : Union[str, Any] = attention_probs_dropout_prob __magic_name__ : Tuple = max_position_embeddings __magic_name__ : str = type_vocab_size __magic_name__ : Dict = initializer_range __magic_name__ : Tuple = layer_norm_eps __magic_name__ : Optional[int] = rotary_value __magic_name__ : List[Any] = use_cache class _snake_case ( snake_case ): @property def SCREAMING_SNAKE_CASE ( self ): if self.task == "multiple-choice": __magic_name__ : str = {0: "batch", 1: "choice", 2: "sequence"} else: __magic_name__ : str = {0: "batch", 1: "sequence"} __magic_name__ : Tuple = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
362
import re import string import numpy as np import datasets snake_case : Any = "\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n" snake_case : Optional[Any] = "\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results[\"exact_match\"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"]\n >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 33.3\n\n" snake_case : Union[str, Any] = "\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , reference_urls=[] , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a=None , _a=False , _a=False , _a=False , ): if regexes_to_ignore is not None: for s in regexes_to_ignore: __magic_name__ : Any = np.array([re.sub(_a , "" , _a ) for x in predictions] ) __magic_name__ : Tuple = np.array([re.sub(_a , "" , _a ) for x in references] ) else: __magic_name__ : Union[str, Any] = np.asarray(_a ) __magic_name__ : List[Any] = np.asarray(_a ) if ignore_case: __magic_name__ : List[Any] = np.char.lower(_a ) __magic_name__ : Optional[int] = np.char.lower(_a ) if ignore_punctuation: __magic_name__ : Optional[Any] = string.punctuation.maketrans("" , "" , string.punctuation ) __magic_name__ : int = np.char.translate(_a , table=_a ) __magic_name__ : Optional[Any] = np.char.translate(_a , table=_a ) if ignore_numbers: __magic_name__ : Optional[Any] = string.digits.maketrans("" , "" , string.digits ) __magic_name__ : Any = np.char.translate(_a , table=_a ) __magic_name__ : List[str] = np.char.translate(_a , table=_a ) __magic_name__ : Dict = predictions == references return {"exact_match": np.mean(_a ) * 100}
41
0
# using dfs for finding eulerian path traversal def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=None ) -> Tuple: '''simple docstring''' lowercase : str = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: lowercase , lowercase : int = True, True lowercase : List[Any] = dfs(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) return path def snake_case( __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' lowercase : List[Any] = 0 lowercase : Union[str, Any] = -1 for i in range(__magic_name__ ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 lowercase : Any = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def snake_case( __magic_name__ , __magic_name__ ) -> Optional[int]: '''simple docstring''' lowercase : Optional[int] = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] lowercase , lowercase : Union[str, Any] = check_circuit_or_path(__magic_name__ , __magic_name__ ) if check == 3: print('''graph is not Eulerian''' ) print('''no path''' ) return lowercase : Optional[Any] = 1 if check == 2: lowercase : Optional[int] = odd_node print('''graph has a Euler path''' ) if check == 1: print('''graph has a Euler cycle''' ) lowercase : int = dfs(__magic_name__ , __magic_name__ , __magic_name__ ) print(__magic_name__ ) def snake_case( ) -> List[str]: '''simple docstring''' lowercase : Tuple = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} lowercase : Optional[Any] = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} lowercase : List[Any] = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} lowercase : str = {1: [2, 3], 2: [1, 3], 3: [1, 2]} lowercase : List[Any] = { 1: [], 2: [] # all degree is zero } lowercase : Union[str, Any] = 10 check_euler(__magic_name__ , __magic_name__ ) check_euler(__magic_name__ , __magic_name__ ) check_euler(__magic_name__ , __magic_name__ ) check_euler(__magic_name__ , __magic_name__ ) check_euler(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
308
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(_lowerCamelCase ) , '''Tatoeba directory does not exist.''' ) class _A ( unittest.TestCase ): @cached_property def __a ( self : int ) -> Dict: """simple docstring""" lowercase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=_A ) @slow def __a ( self : Any ) -> List[Any]: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def __a ( self : int ) -> Tuple: """simple docstring""" lowercase , lowercase : Optional[Any] = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=_A ) assert mmeta["long_pair"] == "heb-eng"
308
1
from argparse import ArgumentParser from .env import EnvironmentCommand def A__ ( ): SCREAMING_SNAKE_CASE_ = ArgumentParser('''Diffusers CLI tool''', usage='''diffusers-cli <command> [<args>]''' ) SCREAMING_SNAKE_CASE_ = parser.add_subparsers(help='''diffusers-cli command helpers''' ) # Register commands EnvironmentCommand.register_subcommand(__lowerCamelCase ) # Let's go SCREAMING_SNAKE_CASE_ = parser.parse_args() if not hasattr(__lowerCamelCase, '''func''' ): parser.print_help() exit(1 ) # Run SCREAMING_SNAKE_CASE_ = args.func(__lowerCamelCase ) service.run() if __name__ == "__main__": main()
257
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { "microsoft/unispeech-sat-base-100h-libri-ft": ( "https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json" ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ ="unispeech-sat" def __init__( self , _A=32 , _A=768 , _A=12 , _A=12 , _A=3072 , _A="gelu" , _A=0.1 , _A=0.1 , _A=0.1 , _A=0.0 , _A=0.0 , _A=0.1 , _A=0.1 , _A=0.02 , _A=1E-5 , _A="group" , _A="gelu" , _A=(512, 512, 512, 512, 512, 512, 512) , _A=(5, 2, 2, 2, 2, 2, 2) , _A=(10, 3, 3, 3, 3, 2, 2) , _A=False , _A=128 , _A=16 , _A=False , _A=True , _A=0.05 , _A=10 , _A=2 , _A=0.0 , _A=10 , _A=0 , _A=320 , _A=2 , _A=0.1 , _A=100 , _A=256 , _A=256 , _A=0.1 , _A="mean" , _A=False , _A=False , _A=256 , _A=(512, 512, 512, 512, 1500) , _A=(5, 3, 3, 1, 1) , _A=(1, 2, 3, 1, 1) , _A=512 , _A=0 , _A=1 , _A=2 , _A=504 , **_A , ) -> Tuple: super().__init__(**_A , pad_token_id=_A , bos_token_id=_A , eos_token_id=_A ) SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = feat_extract_norm SCREAMING_SNAKE_CASE_ = feat_extract_activation SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = conv_bias SCREAMING_SNAKE_CASE_ = num_conv_pos_embeddings SCREAMING_SNAKE_CASE_ = num_conv_pos_embedding_groups SCREAMING_SNAKE_CASE_ = len(self.conv_dim ) SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = hidden_dropout SCREAMING_SNAKE_CASE_ = attention_dropout SCREAMING_SNAKE_CASE_ = activation_dropout SCREAMING_SNAKE_CASE_ = feat_proj_dropout SCREAMING_SNAKE_CASE_ = final_dropout SCREAMING_SNAKE_CASE_ = layerdrop SCREAMING_SNAKE_CASE_ = layer_norm_eps SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = num_clusters SCREAMING_SNAKE_CASE_ = do_stable_layer_norm SCREAMING_SNAKE_CASE_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 SCREAMING_SNAKE_CASE_ = apply_spec_augment SCREAMING_SNAKE_CASE_ = mask_time_prob SCREAMING_SNAKE_CASE_ = mask_time_length SCREAMING_SNAKE_CASE_ = mask_time_min_masks SCREAMING_SNAKE_CASE_ = mask_feature_prob SCREAMING_SNAKE_CASE_ = mask_feature_length SCREAMING_SNAKE_CASE_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations SCREAMING_SNAKE_CASE_ = num_codevectors_per_group SCREAMING_SNAKE_CASE_ = num_codevector_groups SCREAMING_SNAKE_CASE_ = contrastive_logits_temperature SCREAMING_SNAKE_CASE_ = feat_quantizer_dropout SCREAMING_SNAKE_CASE_ = num_negatives SCREAMING_SNAKE_CASE_ = codevector_dim SCREAMING_SNAKE_CASE_ = proj_codevector_dim SCREAMING_SNAKE_CASE_ = diversity_loss_weight # ctc loss SCREAMING_SNAKE_CASE_ = ctc_loss_reduction SCREAMING_SNAKE_CASE_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = xvector_output_dim @property def _UpperCamelCase ( self ) -> str: return functools.reduce(operator.mul , self.conv_stride , 1 )
257
1