code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
101
"""simple docstring""" from collections import defaultdict class lowerCamelCase : def __init__( self : List[str] , __UpperCAmelCase : Dict , __UpperCAmelCase : Any ) -> Any: SCREAMING_SNAKE_CASE__ = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 SCREAMING_SNAKE_CASE__ = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(__UpperCAmelCase ) ) ] SCREAMING_SNAKE_CASE__ = defaultdict(__UpperCAmelCase ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 SCREAMING_SNAKE_CASE__ = (1 << len(__UpperCAmelCase )) - 1 def SCREAMING_SNAKE_CASE ( self : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[int] ) -> Optional[int]: # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement SCREAMING_SNAKE_CASE__ = self.count_ways_until(__UpperCAmelCase , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. SCREAMING_SNAKE_CASE__ = total_ways_util return self.dp[mask][task_no] def SCREAMING_SNAKE_CASE ( self : Optional[int] , __UpperCAmelCase : Union[str, Any] ) -> List[str]: # Store the list of persons for each task for i in range(len(__UpperCAmelCase ) ): for j in task_performed[i]: self.task[j].append(__UpperCAmelCase ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": A_ : Any = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. A_ : Any = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
165
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging A : Optional[int] = logging.get_logger(__name__) class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = '''encoder-decoder''' A__ = True def __init__(self : Union[str, Any] , **_UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" super().__init__(**_UpperCAmelCase ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" lowercase__ = kwargs.pop("""encoder""" ) lowercase__ = encoder_config.pop("""model_type""" ) lowercase__ = kwargs.pop("""decoder""" ) lowercase__ = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig lowercase__ = AutoConfig.for_model(_UpperCAmelCase , **_UpperCAmelCase ) lowercase__ = AutoConfig.for_model(_UpperCAmelCase , **_UpperCAmelCase ) lowercase__ = True @classmethod def lowerCamelCase__ (cls : Union[str, Any] , _UpperCAmelCase : PretrainedConfig , _UpperCAmelCase : PretrainedConfig , **_UpperCAmelCase : int ) -> PretrainedConfig: """simple docstring""" logger.info("""Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) lowercase__ = True lowercase__ = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_UpperCAmelCase ) def lowerCamelCase__ (self : int ) -> Any: """simple docstring""" lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.encoder.to_dict() lowercase__ = self.decoder.to_dict() lowercase__ = self.__class__.model_type return output
146
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor A : List[Any] = logging.get_logger(__name__) class A ( UpperCAmelCase__ ): '''simple docstring''' def __init__(self : List[Any] , *_UpperCAmelCase : int , **_UpperCAmelCase : List[str] ) -> None: """simple docstring""" warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase )
146
1
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class _a : def __init__( self: int , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[str]=14 , UpperCamelCase_: Optional[Any]=7 , UpperCamelCase_: Any=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: str=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Any=True , UpperCamelCase_: Optional[Any]=99 , UpperCamelCase_: int=32 , UpperCamelCase_: Tuple=5 , UpperCamelCase_: str=4 , UpperCamelCase_: str=37 , UpperCamelCase_: Any="gelu" , UpperCamelCase_: int=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: Optional[Any]=512 , UpperCamelCase_: Dict=16 , UpperCamelCase_: Dict=2 , UpperCamelCase_: Any=0.02 , UpperCamelCase_: Any=3 , UpperCamelCase_: Dict=4 , UpperCamelCase_: int=None , ) -> str: """simple docstring""" lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_token_type_ids lowercase__ = use_input_mask lowercase__ = use_labels lowercase__ = use_mc_token_ids lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = num_labels lowercase__ = num_choices lowercase__ = scope lowercase__ = self.vocab_size - 1 def lowerCamelCase_ ( self: Union[str, Any] ) -> str: """simple docstring""" lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ = None if self.use_input_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ = None if self.use_token_type_ids: lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__ = None if self.use_mc_token_ids: lowercase__ = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) lowercase__ = None lowercase__ = None lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase__ = ids_tensor([self.batch_size] , self.num_choices ) lowercase__ = self.get_config() lowercase__ = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase_ ( self: Dict ) -> int: """simple docstring""" return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCamelCase_ ( self: int , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[int] , *UpperCamelCase_: Any ) -> Dict: """simple docstring""" lowercase__ = CTRLModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , head_mask=UpperCamelCase_ ) model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ ) lowercase__ = model(UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Tuple , UpperCamelCase_: Optional[Any] , *UpperCamelCase_: int ) -> int: """simple docstring""" lowercase__ = CTRLLMHeadModel(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() lowercase__ = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase_ ( self: Optional[Any] ) -> Dict: """simple docstring""" lowercase__ = self.prepare_config_and_inputs() ( ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ) = config_and_inputs lowercase__ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask} return config, inputs_dict def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: List[str] , UpperCamelCase_: int , UpperCamelCase_: Any , *UpperCamelCase_: Tuple ) -> int: """simple docstring""" lowercase__ = self.num_labels lowercase__ = CTRLForSequenceClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class _a ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): _lowercase : Union[str, Any] = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () _lowercase : Optional[int] = (CTRLLMHeadModel,) if is_torch_available() else () _lowercase : str = ( { '''feature-extraction''': CTRLModel, '''text-classification''': CTRLForSequenceClassification, '''text-generation''': CTRLLMHeadModel, '''zero-shot''': CTRLForSequenceClassification, } if is_torch_available() else {} ) _lowercase : List[str] = True _lowercase : Tuple = False _lowercase : Optional[int] = False def lowerCamelCase_ ( self: Tuple , UpperCamelCase_: List[Any] , UpperCamelCase_: int , UpperCamelCase_: Any , UpperCamelCase_: Any , UpperCamelCase_: Tuple ) -> List[Any]: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCamelCase_ ( self: Dict ) -> str: """simple docstring""" lowercase__ = CTRLModelTester(self ) lowercase__ = ConfigTester(self , config_class=UpperCamelCase_ , n_embd=37 ) def lowerCamelCase_ ( self: Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self: int ) -> int: """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase_ ( self: Optional[Any] ) -> List[Any]: """simple docstring""" lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*UpperCamelCase_ ) def lowerCamelCase_ ( self: Tuple ) -> str: """simple docstring""" lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*UpperCamelCase_ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def lowerCamelCase_ ( self: Dict ) -> str: """simple docstring""" pass @slow def lowerCamelCase_ ( self: List[str] ) -> List[Any]: """simple docstring""" for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = CTRLModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def lowerCamelCase_ ( self: List[str] ) -> Tuple: """simple docstring""" pass @require_torch class _a ( unittest.TestCase ): def lowerCamelCase_ ( self: List[str] ) -> List[str]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCamelCase_ ( self: Optional[Any] ) -> str: """simple docstring""" lowercase__ = CTRLLMHeadModel.from_pretrained('''ctrl''' ) model.to(UpperCamelCase_ ) lowercase__ = torch.tensor( [[11_859, 0, 1_611, 8]] , dtype=torch.long , device=UpperCamelCase_ ) # Legal the president is lowercase__ = [ 11_859, 0, 1_611, 8, 5, 150, 26_449, 2, 19, 348, 469, 3, 2_595, 48, 20_740, 246_533, 246_533, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a lowercase__ = model.generate(UpperCamelCase_ , do_sample=UpperCamelCase_ ) self.assertListEqual(output_ids[0].tolist() , UpperCamelCase_ )
110
class _a : def __init__( self: Any ) -> Tuple: """simple docstring""" lowercase__ = '''''' lowercase__ = '''''' lowercase__ = [] def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int ) -> int: """simple docstring""" if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: lowercase__ = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: lowercase__ = self.__min_dist_top_down_dp(UpperCamelCase_ , n - 1 ) lowercase__ = self.__min_dist_top_down_dp(m - 1 , UpperCamelCase_ ) lowercase__ = self.__min_dist_top_down_dp(m - 1 , n - 1 ) lowercase__ = 1 + min(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) return self.dp[m][n] def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: str ) -> int: """simple docstring""" lowercase__ = worda lowercase__ = worda lowercase__ = [[-1 for _ in range(len(UpperCamelCase_ ) )] for _ in range(len(UpperCamelCase_ ) )] return self.__min_dist_top_down_dp(len(UpperCamelCase_ ) - 1 , len(UpperCamelCase_ ) - 1 ) def lowerCamelCase_ ( self: int , UpperCamelCase_: str , UpperCamelCase_: str ) -> int: """simple docstring""" lowercase__ = worda lowercase__ = worda lowercase__ = len(UpperCamelCase_ ) lowercase__ = len(UpperCamelCase_ ) lowercase__ = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty lowercase__ = j elif j == 0: # second string is empty lowercase__ = i elif worda[i - 1] == worda[j - 1]: # last characters are equal lowercase__ = self.dp[i - 1][j - 1] else: lowercase__ = self.dp[i][j - 1] lowercase__ = self.dp[i - 1][j] lowercase__ = self.dp[i - 1][j - 1] lowercase__ = 1 + min(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) return self.dp[m][n] if __name__ == "__main__": lowerCAmelCase = EditDistance() print('****************** Testing Edit Distance DP Algorithm ******************') print() lowerCAmelCase = input('Enter the first string: ').strip() lowerCAmelCase = input('Enter the second string: ').strip() print() print(f"""The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}""") print(f"""The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}""") print() print('*************** End of Testing Edit Distance DP Algorithm ***************')
110
1
"""simple docstring""" from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def __a ( __lowerCamelCase ) -> int: UpperCAmelCase_ : int = int(number**0.5 ) return number == sq * sq def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) -> Optional[int]: UpperCAmelCase_ : int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den UpperCAmelCase_ : int = x_den * y_den * z_den UpperCAmelCase_ : int = gcd(__lowerCamelCase, __lowerCamelCase ) top //= hcf bottom //= hcf return top, bottom def __a ( __lowerCamelCase = 35 ) -> int: UpperCAmelCase_ : set = set() UpperCAmelCase_ : int UpperCAmelCase_ : Fraction = Fraction(0 ) UpperCAmelCase_ : tuple[int, int] for x_num in range(1, order + 1 ): for x_den in range(x_num + 1, order + 1 ): for y_num in range(1, order + 1 ): for y_den in range(y_num + 1, order + 1 ): # n=1 UpperCAmelCase_ : Union[str, Any] = x_num * y_den + x_den * y_num UpperCAmelCase_ : List[str] = x_den * y_den UpperCAmelCase_ : Tuple = gcd(__lowerCamelCase, __lowerCamelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ : Union[str, Any] = add_three( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) unique_s.add(__lowerCamelCase ) # n=2 UpperCAmelCase_ : Union[str, Any] = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) UpperCAmelCase_ : int = x_den * x_den * y_den * y_den if is_sq(__lowerCamelCase ) and is_sq(__lowerCamelCase ): UpperCAmelCase_ : List[Any] = int(sqrt(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = int(sqrt(__lowerCamelCase ) ) UpperCAmelCase_ : Tuple = gcd(__lowerCamelCase, __lowerCamelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ : Tuple = add_three( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) unique_s.add(__lowerCamelCase ) # n=-1 UpperCAmelCase_ : Any = x_num * y_num UpperCAmelCase_ : Dict = x_den * y_num + x_num * y_den UpperCAmelCase_ : Any = gcd(__lowerCamelCase, __lowerCamelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ : List[Any] = add_three( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) unique_s.add(__lowerCamelCase ) # n=2 UpperCAmelCase_ : Optional[Any] = x_num * x_num * y_num * y_num UpperCAmelCase_ : Union[str, Any] = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(__lowerCamelCase ) and is_sq(__lowerCamelCase ): UpperCAmelCase_ : Tuple = int(sqrt(__lowerCamelCase ) ) UpperCAmelCase_ : List[Any] = int(sqrt(__lowerCamelCase ) ) UpperCAmelCase_ : List[Any] = gcd(__lowerCamelCase, __lowerCamelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ : List[str] = add_three( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) unique_s.add(__lowerCamelCase ) for num, den in unique_s: total += Fraction(__lowerCamelCase, __lowerCamelCase ) return total.denominator + total.numerator if __name__ == "__main__": print(f"""{solution() = }""")
371
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
0
from __future__ import annotations def lowerCamelCase__ ( snake_case_ : list[int] , snake_case_ : list[int] , snake_case_ : int ) -> tuple[float, list[float]]: __snake_case = list(range(len(snake_case_ ) ) ) __snake_case = [v / w for v, w in zip(snake_case_ , snake_case_ )] index.sort(key=lambda snake_case_ : ratio[i] , reverse=snake_case_ ) __snake_case = 0 __snake_case = [0] * len(snake_case_ ) for i in index: if weight[i] <= capacity: __snake_case = 1 max_value += value[i] capacity -= weight[i] else: __snake_case = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
24
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : def __init__(self : Any , a__ : Union[str, Any] , a__ : int=13 , a__ : int=7 , a__ : Optional[Any]=True , a__ : Optional[int]=True , a__ : Any=True , a__ : str=True , a__ : List[Any]=99 , a__ : Any=24 , a__ : List[str]=2 , a__ : Optional[int]=6 , a__ : int=37 , a__ : List[str]="gelu" , a__ : List[Any]=0.1 , a__ : Optional[int]=0.1 , a__ : Union[str, Any]=512 , a__ : List[str]=16 , a__ : Optional[int]=2 , a__ : Union[str, Any]=0.0_2 , a__ : str=3 , a__ : Optional[Any]=None , a__ : Any=1000 , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = scope __snake_case = range_bbox def a (self : Optional[int] ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __snake_case = bbox[i, j, 3] __snake_case = bbox[i, j, 1] __snake_case = t if bbox[i, j, 2] < bbox[i, j, 0]: __snake_case = bbox[i, j, 2] __snake_case = bbox[i, j, 0] __snake_case = t __snake_case = None if self.use_input_mask: __snake_case = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __snake_case = None if self.use_token_type_ids: __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def a (self : List[str] ): """simple docstring""" return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def a (self : List[Any] , a__ : List[Any] , a__ : Optional[Any] , a__ : List[str] , a__ : int , a__ : Optional[int] , a__ : str , a__ : Optional[int] , ): """simple docstring""" __snake_case = LiltModel(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ ) __snake_case = model(a__ , bbox=a__ , token_type_ids=a__ ) __snake_case = model(a__ , bbox=a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a (self : Any , a__ : Tuple , a__ : Dict , a__ : Optional[int] , a__ : Dict , a__ : Union[str, Any] , a__ : str , a__ : Tuple , ): """simple docstring""" __snake_case = self.num_labels __snake_case = LiltForTokenClassification(config=a__ ) model.to(a__ ) model.eval() __snake_case = model( a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a (self : int , a__ : Optional[Any] , a__ : int , a__ : int , a__ : Optional[Any] , a__ : Tuple , a__ : Union[str, Any] , a__ : str , ): """simple docstring""" __snake_case = LiltForQuestionAnswering(config=a__ ) model.to(a__ ) model.eval() __snake_case = model( a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ , start_positions=a__ , end_positions=a__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a (self : Tuple ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A_ : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) A_ : Any = ( { 'feature-extraction': LiltModel, 'question-answering': LiltForQuestionAnswering, 'text-classification': LiltForSequenceClassification, 'token-classification': LiltForTokenClassification, 'zero-shot': LiltForSequenceClassification, } if is_torch_available() else {} ) A_ : Optional[int] = False A_ : List[Any] = False def a (self : Dict , a__ : Tuple , a__ : Tuple , a__ : Tuple , a__ : Union[str, Any] , a__ : Any ): """simple docstring""" return True def a (self : Union[str, Any] ): """simple docstring""" __snake_case = LiltModelTester(self ) __snake_case = ConfigTester(self , config_class=a__ , hidden_size=37 ) def a (self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() def a (self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def a (self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __snake_case = type self.model_tester.create_and_check_model(*a__ ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a__ ) def a (self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a__ ) @slow def a (self : Optional[int] ): """simple docstring""" for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = LiltModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) @require_torch @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def a (self : Tuple ): """simple docstring""" __snake_case = LiltModel.from_pretrained('''SCUT-DLVCLab/lilt-roberta-en-base''' ).to(a__ ) __snake_case = torch.tensor([[1, 2]] , device=a__ ) __snake_case = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=a__ ) # forward pass with torch.no_grad(): __snake_case = model(input_ids=a__ , bbox=a__ ) __snake_case = torch.Size([1, 2, 768] ) __snake_case = torch.tensor( [[-0.0_6_5_3, 0.0_9_5_0, -0.0_0_6_1], [-0.0_5_4_5, 0.0_9_2_6, -0.0_3_2_4]] , device=a__ , ) self.assertTrue(outputs.last_hidden_state.shape , a__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , a__ , atol=1E-3 ) )
24
1
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ): lowerCAmelCase__ = [randint(-1_0_0_0 , 1_0_0_0 ) for i in range(1_0 )] lowerCAmelCase__ = randint(-5_0_0_0 , 5_0_0_0 ) return (arr, r) lowerCAmelCase__ = make_dataset() def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): for triplet in permutations(lowerCAmelCase__ , 3 ): if sum(lowerCAmelCase__ ) == target: return tuple(sorted(lowerCAmelCase__ ) ) return (0, 0, 0) def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): arr.sort() lowerCAmelCase__ = len(lowerCAmelCase__ ) for i in range(n - 1 ): lowerCAmelCase__ , lowerCAmelCase__ = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ): lowerCAmelCase__ = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' lowerCAmelCase__ = '\ntriplet_sum1(*dataset)\n' lowerCAmelCase__ = '\ntriplet_sum2(*dataset)\n' lowerCAmelCase__ = repeat(setup=lowerCAmelCase__ , stmt=lowerCAmelCase__ , repeat=5 , number=1_0_0_0_0 ) lowerCAmelCase__ = repeat(setup=lowerCAmelCase__ , stmt=lowerCAmelCase__ , repeat=5 , number=1_0_0_0_0 ) return (min(lowerCAmelCase__ ), min(lowerCAmelCase__ )) if __name__ == "__main__": from doctest import testmod testmod() lowerCAmelCase__ = solution_times() print(F'The time for naive implementation is {times[0]}.') print(F'The time for optimized implementation is {times[1]}.')
119
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class a_ : '''simple docstring''' UpperCAmelCase_ = None UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = True UpperCAmelCase_ = None UpperCAmelCase_ = 1 UpperCAmelCase_ = None UpperCAmelCase_ = False UpperCAmelCase_ = None UpperCAmelCase_ = None def __snake_case ( self : Dict): '''simple docstring''' return self.__class__(**{k: copy.deepcopy(lowercase__) for k, v in self.__dict__.items()})
119
1
'''simple docstring''' import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) lowerCamelCase :int = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(3_2, (3, 3), input_shape=(6_4, 6_4, 3), activation='''relu''') ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(3_2, (3, 3), activation='''relu''')) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=1_2_8, activation='''relu''')) classifier.add(layers.Dense(units=1, activation='''sigmoid''')) # Compiling the CNN classifier.compile( optimizer='''adam''', loss='''binary_crossentropy''', metrics=['''accuracy'''] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') lowerCamelCase :str = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 2_5_5, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) lowerCamelCase :Tuple = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 2_5_5) lowerCamelCase :Any = train_datagen.flow_from_directory( '''dataset/training_set''', target_size=(6_4, 6_4), batch_size=3_2, class_mode='''binary''' ) lowerCamelCase :Tuple = test_datagen.flow_from_directory( '''dataset/test_set''', target_size=(6_4, 6_4), batch_size=3_2, class_mode='''binary''' ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=3_0, validation_data=test_set ) classifier.save('''cnn.h5''') # Part 3 - Making new predictions lowerCamelCase :int = tf.keras.preprocessing.image.load_img( '''dataset/single_prediction/image.png''', target_size=(6_4, 6_4) ) lowerCamelCase :Optional[int] = tf.keras.preprocessing.image.img_to_array(test_image) lowerCamelCase :Tuple = np.expand_dims(test_image, axis=0) lowerCamelCase :int = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: lowerCamelCase :List[Any] = 'Normal' if result[0][0] == 1: lowerCamelCase :Tuple = 'Abnormality detected'
206
"""simple docstring""" def lowercase__ ( _UpperCAmelCase ) -> None: '''simple docstring''' lowercase : Union[str, Any] = generate_pascal_triangle(_UpperCAmelCase ) for row_idx in range(_UpperCAmelCase ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=' ' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=' ' ) else: print(triangle[row_idx][col_idx] , end='' ) print() def lowercase__ ( _UpperCAmelCase ) -> list[list[int]]: '''simple docstring''' if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) lowercase : list[list[int]] = [] for current_row_idx in range(_UpperCAmelCase ): lowercase : Optional[Any] = populate_current_row(_UpperCAmelCase , _UpperCAmelCase ) triangle.append(_UpperCAmelCase ) return triangle def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase ) -> list[int]: '''simple docstring''' lowercase : Union[str, Any] = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 lowercase , lowercase : List[str] = 1, 1 for current_col_idx in range(1 , _UpperCAmelCase ): calculate_current_element( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return current_row def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> None: '''simple docstring''' lowercase : List[str] = triangle[current_row_idx - 1][current_col_idx - 1] lowercase : Optional[Any] = triangle[current_row_idx - 1][current_col_idx] lowercase : Any = above_to_left_elt + above_to_right_elt def lowercase__ ( _UpperCAmelCase ) -> list[list[int]]: '''simple docstring''' if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) lowercase : list[list[int]] = [[1]] for row_index in range(1 , _UpperCAmelCase ): lowercase : List[str] = [0] + result[-1] + [0] lowercase : List[Any] = row_index + 1 # Calculate the number of distinct elements in a row lowercase : Union[str, Any] = sum(divmod(_UpperCAmelCase , 2 ) ) lowercase : Tuple = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] lowercase : str = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() lowercase : List[Any] = row_first_half + row_second_half result.append(_UpperCAmelCase ) return result def lowercase__ ( ) -> None: '''simple docstring''' from collections.abc import Callable from timeit import timeit def benchmark_a_function(_UpperCAmelCase , _UpperCAmelCase ) -> None: lowercase : Dict = f'''{func.__name__}({value})''' lowercase : Optional[Any] = timeit(f'''__main__.{call}''' , setup='import __main__' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f'''{call:38} -- {timing:.4f} seconds''' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(_UpperCAmelCase , _UpperCAmelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
255
0
import collections import os import re from pathlib import Path _UpperCAmelCase = 'src/transformers' # Matches is_xxx_available() _UpperCAmelCase = re.compile(r'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} _UpperCAmelCase = re.compile(r'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] _UpperCAmelCase = re.compile(r'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available _UpperCAmelCase = re.compile(r'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") _UpperCAmelCase = re.compile(r'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] _UpperCAmelCase = re.compile(r'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", _UpperCAmelCase = re.compile(r'^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], _UpperCAmelCase = re.compile(r'^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo _UpperCAmelCase = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: _UpperCAmelCase = re.compile(r'^\s*try:') # Catches a line with else: _UpperCAmelCase = re.compile(r'^\s*else:') def lowerCAmelCase_ ( UpperCamelCase_ ) -> Any: if _re_test_backend.search(UpperCamelCase_ ) is None: return None UpperCamelCase_ = [b[0] for b in _re_backend.findall(UpperCamelCase_ )] backends.sort() return "_and_".join(UpperCamelCase_ ) def lowerCAmelCase_ ( UpperCamelCase_ ) -> Optional[int]: with open(UpperCamelCase_ , "r" , encoding="utf-8" , newline="\n" ) as f: UpperCamelCase_ = f.readlines() UpperCamelCase_ = 0 while line_index < len(UpperCamelCase_ ) and not lines[line_index].startswith("_import_structure = {" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(UpperCamelCase_ ): return None # First grab the objects without a specific backend in _import_structure UpperCamelCase_ = [] while not lines[line_index].startswith("if TYPE_CHECKING" ) and find_backend(lines[line_index] ) is None: UpperCamelCase_ = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(UpperCamelCase_ ): UpperCamelCase_ = _re_one_line_import_struct.search(UpperCamelCase_ ).groups()[0] UpperCamelCase_ = re.findall(r"\[([^\]]+)\]" , UpperCamelCase_ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(", " )] ) line_index += 1 continue UpperCamelCase_ = _re_import_struct_key_value.search(UpperCamelCase_ ) if single_line_import_search is not None: UpperCamelCase_ = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(", " ) if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) line_index += 1 UpperCamelCase_ = {"none": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("if TYPE_CHECKING" ): # If the line is an if not is_backend_available, we grab all objects associated. UpperCamelCase_ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: UpperCamelCase_ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 UpperCamelCase_ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 4 ): UpperCamelCase_ = lines[line_index] if _re_import_struct_add_one.search(UpperCamelCase_ ) is not None: objects.append(_re_import_struct_add_one.search(UpperCamelCase_ ).groups()[0] ) elif _re_import_struct_add_many.search(UpperCamelCase_ ) is not None: UpperCamelCase_ = _re_import_struct_add_many.search(UpperCamelCase_ ).groups()[0].split(", " ) UpperCamelCase_ = [obj[1:-1] for obj in imports if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif _re_between_brackets.search(UpperCamelCase_ ) is not None: UpperCamelCase_ = _re_between_brackets.search(UpperCamelCase_ ).groups()[0].split(", " ) UpperCamelCase_ = [obj[1:-1] for obj in imports if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif _re_quote_object.search(UpperCamelCase_ ) is not None: objects.append(_re_quote_object.search(UpperCamelCase_ ).groups()[0] ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) elif line.startswith(" " * 12 + "\"" ): objects.append(line[13:-3] ) line_index += 1 UpperCamelCase_ = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend UpperCamelCase_ = [] while ( line_index < len(UpperCamelCase_ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("else" ) ): UpperCamelCase_ = lines[line_index] UpperCamelCase_ = _re_import.search(UpperCamelCase_ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 UpperCamelCase_ = {"none": objects} # Let's continue with backend-specific objects while line_index < len(UpperCamelCase_ ): # If the line is an if is_backend_available, we grab all objects associated. UpperCamelCase_ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: UpperCamelCase_ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 UpperCamelCase_ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 8 ): UpperCamelCase_ = lines[line_index] UpperCamelCase_ = _re_import.search(UpperCamelCase_ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 12 ): objects.append(line[12:-2] ) line_index += 1 UpperCamelCase_ = objects else: line_index += 1 return import_dict_objects, type_hint_objects def lowerCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[Any]: def find_duplicates(UpperCamelCase_ ): return [k for k, v in collections.Counter(UpperCamelCase_ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] UpperCamelCase_ = [] for key in import_dict_objects.keys(): UpperCamelCase_ = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) UpperCamelCase_ = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): UpperCamelCase_ = "base imports" if key == "none" else F'''{key} backend''' errors.append(F'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def lowerCAmelCase_ ( ) -> str: UpperCamelCase_ = [] for root, _, files in os.walk(UpperCamelCase_ ): if "__init__.py" in files: UpperCamelCase_ = os.path.join(UpperCamelCase_ , "__init__.py" ) UpperCamelCase_ = parse_init(UpperCamelCase_ ) if objects is not None: UpperCamelCase_ = analyze_results(*UpperCamelCase_ ) if len(UpperCamelCase_ ) > 0: UpperCamelCase_ = F'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append("\n".join(UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > 0: raise ValueError("\n\n".join(UpperCamelCase_ ) ) def lowerCAmelCase_ ( ) -> List[Any]: UpperCamelCase_ = [] for path, directories, files in os.walk(UpperCamelCase_ ): for folder in directories: # Ignore private modules if folder.startswith("_" ): directories.remove(UpperCamelCase_ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(UpperCamelCase_ ) / folder).glob("*.py" ) ) ) == 0: continue UpperCamelCase_ = str((Path(UpperCamelCase_ ) / folder).relative_to(UpperCamelCase_ ) ) UpperCamelCase_ = short_path.replace(os.path.sep , "." ) submodules.append(UpperCamelCase_ ) for fname in files: if fname == "__init__.py": continue UpperCamelCase_ = str((Path(UpperCamelCase_ ) / fname).relative_to(UpperCamelCase_ ) ) UpperCamelCase_ = short_path.replace(".py" , "" ).replace(os.path.sep , "." ) if len(submodule.split("." ) ) == 1: submodules.append(UpperCamelCase_ ) return submodules _UpperCAmelCase = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', 'models.esm.openfold_utils', ] def lowerCAmelCase_ ( ) -> List[str]: # This is to make sure the transformers module imported is the one in the repo. from transformers.utils import direct_transformers_import UpperCamelCase_ = direct_transformers_import(UpperCamelCase_ ) UpperCamelCase_ = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(UpperCamelCase_ , "__init__.py" ) , "r" ) as f: UpperCamelCase_ = f.read() import_structure_keys.update(set(re.findall(r"import_structure\[\"([^\"]*)\"\]" , UpperCamelCase_ ) ) ) UpperCamelCase_ = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(UpperCamelCase_ ) > 0: UpperCamelCase_ = "\n".join(F'''- {module}''' for module in module_not_registered ) raise ValueError( "The following submodules are not properly registed in the main init of Transformers:\n" F'''{list_of_modules}\n''' "Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value." ) if __name__ == "__main__": check_all_inits() check_submodules()
328
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _UpperCamelCase ( lowerCAmelCase_ , unittest.TestCase ): _UpperCamelCase : Optional[Any] = DiTPipeline _UpperCamelCase : Any = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS _UpperCamelCase : Dict = PipelineTesterMixin.required_optional_params - { '''latents''', '''num_images_per_prompt''', '''callback''', '''callback_steps''', } _UpperCamelCase : Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS _UpperCamelCase : Dict = False def lowercase ( self: str ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) UpperCamelCase_ = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=_SCREAMING_SNAKE_CASE , activation_fn="gelu-approximate" , num_embeds_ada_norm=1000 , norm_type="ada_norm_zero" , norm_elementwise_affine=_SCREAMING_SNAKE_CASE , ) UpperCamelCase_ = AutoencoderKL() UpperCamelCase_ = DDIMScheduler() UpperCamelCase_ = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def lowercase ( self: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=0 ) -> Dict: """simple docstring""" if str(_SCREAMING_SNAKE_CASE ).startswith("mps" ): UpperCamelCase_ = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: UpperCamelCase_ = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def lowercase ( self: Any ) -> List[str]: """simple docstring""" UpperCamelCase_ = "cpu" UpperCamelCase_ = self.get_dummy_components() UpperCamelCase_ = self.pipeline_class(**_SCREAMING_SNAKE_CASE ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = pipe(**_SCREAMING_SNAKE_CASE ).images UpperCamelCase_ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) UpperCamelCase_ = np.array([0.29_46, 0.66_01, 0.43_29, 0.32_96, 0.41_44, 0.53_19, 0.72_73, 0.50_13, 0.44_57] ) UpperCamelCase_ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3 ) def lowercase ( self: Optional[int] ) -> Any: """simple docstring""" self._test_inference_batch_single_identical(relax_max_difference=_SCREAMING_SNAKE_CASE , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowercase ( self: Optional[Any] ) -> Optional[int]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class _UpperCamelCase ( unittest.TestCase ): def lowercase ( self: Optional[int] ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self: Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) UpperCamelCase_ = ["vase", "umbrella", "white shark", "white wolf"] UpperCamelCase_ = pipe.get_label_ids(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = pipe(_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , num_inference_steps=40 , output_type="np" ).images for word, image in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase_ = load_numpy( f'''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-2 def lowercase ( self: int ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) UpperCamelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) UpperCamelCase_ = ["vase", "umbrella"] UpperCamelCase_ = pipe.get_label_ids(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe(_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , num_inference_steps=25 , output_type="np" ).images for word, image in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" f'''/dit/{word}_512.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-1
328
1
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _snake_case ( _snake_case , _snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE__ = StableDiffusionDiffEditPipeline SCREAMING_SNAKE_CASE__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'height', 'width', 'image'} | {'image_latents'} SCREAMING_SNAKE_CASE__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {'image'} | {'image_latents'} SCREAMING_SNAKE_CASE__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess SCREAMING_SNAKE_CASE__ = frozenset([] ) def SCREAMING_SNAKE_CASE__ ( self ): torch.manual_seed(0 ) a :Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=_lowerCamelCase , ) a :int = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_lowerCamelCase , set_alpha_to_one=_lowerCamelCase , ) a :Tuple = DDIMInverseScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_lowerCamelCase , set_alpha_to_zero=_lowerCamelCase , ) torch.manual_seed(0 ) a :Dict = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) a :int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) a :Optional[int] = CLIPTextModel(_lowerCamelCase ) a :str = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) a :Optional[Any] = { '''unet''': unet, '''scheduler''': scheduler, '''inverse_scheduler''': inverse_scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase=0 ): a :Optional[Any] = floats_tensor((1, 16, 16) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) a :Tuple = floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) if str(_lowerCamelCase ).startswith('''mps''' ): a :int = torch.manual_seed(_lowerCamelCase ) else: a :int = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) a :Dict = { '''prompt''': '''a dog and a newt''', '''mask_image''': mask, '''image_latents''': latents, '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase=0 ): a :str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) a :str = image.cpu().permute(0 , 2 , 3 , 1 )[0] a :Optional[Any] = Image.fromarray(np.uinta(_lowerCamelCase ) ).convert('''RGB''' ) if str(_lowerCamelCase ).startswith('''mps''' ): a :Dict = torch.manual_seed(_lowerCamelCase ) else: a :Any = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) a :Union[str, Any] = { '''image''': image, '''source_prompt''': '''a cat and a frog''', '''target_prompt''': '''a dog and a newt''', '''generator''': generator, '''num_inference_steps''': 2, '''num_maps_per_mask''': 2, '''mask_encode_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase=0 ): a :List[str] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) a :int = image.cpu().permute(0 , 2 , 3 , 1 )[0] a :Optional[Any] = Image.fromarray(np.uinta(_lowerCamelCase ) ).convert('''RGB''' ) if str(_lowerCamelCase ).startswith('''mps''' ): a :Dict = torch.manual_seed(_lowerCamelCase ) else: a :Any = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) a :List[Any] = { '''image''': image, '''prompt''': '''a cat and a frog''', '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''decode_latents''': True, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self ): if not hasattr(self.pipeline_class , '''_optional_components''' ): return a :Any = self.get_dummy_components() a :str = self.pipeline_class(**_lowerCamelCase ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) a :Union[str, Any] = self.get_dummy_inputs(_lowerCamelCase ) a :List[Any] = pipe(**_lowerCamelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowerCamelCase ) a :int = self.pipeline_class.from_pretrained(_lowerCamelCase ) pipe_loaded.to(_lowerCamelCase ) pipe_loaded.set_progress_bar_config(disable=_lowerCamelCase ) for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowerCamelCase , _lowerCamelCase ) is None , F'''`{optional_component}` did not stay set to None after loading.''' , ) a :Dict = self.get_dummy_inputs(_lowerCamelCase ) a :Tuple = pipe_loaded(**_lowerCamelCase )[0] a :List[str] = np.abs(output - output_loaded ).max() self.assertLess(_lowerCamelCase , 1e-4 ) def SCREAMING_SNAKE_CASE__ ( self ): a :Any = '''cpu''' a :Optional[int] = self.get_dummy_components() a :List[str] = self.pipeline_class(**_lowerCamelCase ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) a :Any = self.get_dummy_mask_inputs(_lowerCamelCase ) a :str = pipe.generate_mask(**_lowerCamelCase ) a :List[str] = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) a :List[Any] = np.array([0] * 9 ) a :str = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowerCamelCase , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def SCREAMING_SNAKE_CASE__ ( self ): a :Any = '''cpu''' a :List[str] = self.get_dummy_components() a :List[Any] = self.pipeline_class(**_lowerCamelCase ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) a :str = self.get_dummy_inversion_inputs(_lowerCamelCase ) a :Any = pipe.invert(**_lowerCamelCase ).images a :Union[str, Any] = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) a :str = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) a :str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowerCamelCase , 1e-3 ) def SCREAMING_SNAKE_CASE__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def SCREAMING_SNAKE_CASE__ ( self ): a :str = '''cpu''' a :str = self.get_dummy_components() a :Union[str, Any] = {'''beta_start''': 0.0_0085, '''beta_end''': 0.012, '''beta_schedule''': '''scaled_linear'''} a :Dict = DPMSolverMultistepScheduler(**_lowerCamelCase ) a :List[str] = DPMSolverMultistepInverseScheduler(**_lowerCamelCase ) a :int = self.pipeline_class(**_lowerCamelCase ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) a :str = self.get_dummy_inversion_inputs(_lowerCamelCase ) a :Tuple = pipe.invert(**_lowerCamelCase ).images a :Any = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) a :int = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) a :Tuple = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowerCamelCase , 1e-3 ) @require_torch_gpu @slow class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def SCREAMING_SNAKE_CASE__ ( cls ): a :Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png''' ) a :Union[str, Any] = raw_image.convert('''RGB''' ).resize((768, 768) ) a :List[Any] = raw_image def SCREAMING_SNAKE_CASE__ ( self ): a :Tuple = torch.manual_seed(0 ) a :int = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=_lowerCamelCase , torch_dtype=torch.floataa ) a :Optional[int] = DDIMScheduler.from_config(pipe.scheduler.config ) a :Optional[int] = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_lowerCamelCase ) a :Optional[Any] = '''a bowl of fruit''' a :Any = '''a bowl of pears''' a :Any = pipe.generate_mask( image=self.raw_image , source_prompt=_lowerCamelCase , target_prompt=_lowerCamelCase , generator=_lowerCamelCase , ) a :Dict = pipe.invert( prompt=_lowerCamelCase , image=self.raw_image , inpaint_strength=0.7 , generator=_lowerCamelCase ).latents a :List[str] = pipe( prompt=_lowerCamelCase , mask_image=_lowerCamelCase , image_latents=_lowerCamelCase , generator=_lowerCamelCase , negative_prompt=_lowerCamelCase , inpaint_strength=0.7 , output_type='''numpy''' , ).images[0] a :List[str] = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1 def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[int] = torch.manual_seed(0 ) a :List[Any] = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=_lowerCamelCase , torch_dtype=torch.floataa ) a :Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) a :Union[str, Any] = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_lowerCamelCase ) a :Dict = '''a bowl of fruit''' a :Optional[Any] = '''a bowl of pears''' a :Tuple = pipe.generate_mask( image=self.raw_image , source_prompt=_lowerCamelCase , target_prompt=_lowerCamelCase , generator=_lowerCamelCase , ) a :Dict = pipe.invert( prompt=_lowerCamelCase , image=self.raw_image , inpaint_strength=0.7 , generator=_lowerCamelCase , num_inference_steps=25 , ).latents a :str = pipe( prompt=_lowerCamelCase , mask_image=_lowerCamelCase , image_latents=_lowerCamelCase , generator=_lowerCamelCase , negative_prompt=_lowerCamelCase , inpaint_strength=0.7 , num_inference_steps=25 , output_type='''numpy''' , ).images[0] a :List[Any] = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1
94
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { '''facebook/xlm-roberta-xl''': '''https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json''', '''facebook/xlm-roberta-xxl''': '''https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json''', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : int = '''xlm-roberta-xl''' def __init__( self ,__UpperCAmelCase=25_0880 ,__UpperCAmelCase=2560 ,__UpperCAmelCase=36 ,__UpperCAmelCase=32 ,__UpperCAmelCase=1_0240 ,__UpperCAmelCase="gelu" ,__UpperCAmelCase=0.1 ,__UpperCAmelCase=0.1 ,__UpperCAmelCase=514 ,__UpperCAmelCase=1 ,__UpperCAmelCase=0.0_2 ,__UpperCAmelCase=1E-05 ,__UpperCAmelCase=1 ,__UpperCAmelCase=0 ,__UpperCAmelCase=2 ,__UpperCAmelCase="absolute" ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,**__UpperCAmelCase ,) -> str: super().__init__(pad_token_id=__UpperCAmelCase ,bos_token_id=__UpperCAmelCase ,eos_token_id=__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : int = num_hidden_layers lowerCAmelCase__ : str = num_attention_heads lowerCAmelCase__ : int = hidden_act lowerCAmelCase__ : Dict = intermediate_size lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : str = attention_probs_dropout_prob lowerCAmelCase__ : Optional[int] = max_position_embeddings lowerCAmelCase__ : List[str] = type_vocab_size lowerCAmelCase__ : List[Any] = initializer_range lowerCAmelCase__ : Tuple = layer_norm_eps lowerCAmelCase__ : int = position_embedding_type lowerCAmelCase__ : Optional[Any] = use_cache lowerCAmelCase__ : Optional[Any] = classifier_dropout class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase__ : Dict = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCAmelCase__ : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
37
0
import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging _lowerCAmelCase : str = logging.get_logger(__name__) def UpperCamelCase_( _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : str , _snake_case : Any=False ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise if not is_sharded: __a =os.path.abspath(_snake_case ) logger.info(F'Loading PyTorch weights from {pt_path}' ) __a =torch.load(_snake_case , map_location='cpu' ) logger.info(F'PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.' ) __a =convert_pytorch_state_dict_to_flax(_snake_case , _snake_case ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files __a =convert_pytorch_sharded_state_dict_to_flax(_snake_case , _snake_case ) return flax_state_dict def UpperCamelCase_( _snake_case : Tuple[str] , _snake_case : np.ndarray , _snake_case : Dict[str, jnp.ndarray] , _snake_case : str , ): """simple docstring""" def is_key_or_prefix_key_in_dict(_snake_case : Tuple[str] ) -> bool: return len(set(_snake_case ) & {key, (model_prefix,) + key} ) > 0 # layer norm __a =pt_tuple_key[:-1] + ('scale',) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(_snake_case ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean __a =pt_tuple_key[:-1] + ('mean',) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(_snake_case ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var __a =pt_tuple_key[:-1] + ('var',) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(_snake_case ): return renamed_pt_tuple_key, pt_tensor # embedding __a =pt_tuple_key[:-1] + ('embedding',) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(_snake_case ): return renamed_pt_tuple_key, pt_tensor # conv layer __a =pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(_snake_case ): __a =pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __a =pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(_snake_case ): __a =pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __a =pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __a =pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 __a =None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): __a =pt_tuple_key[-2] + '_g' elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): __a =pt_tuple_key[-2] + '_v' if name is not None: __a =pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCamelCase_( _snake_case : Dict , _snake_case : List[Any] ): """simple docstring""" __a ={k: v.numpy() for k, v in pt_state_dict.items()} __a =flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: __a =flax_model.params['params'] else: __a =flax_model.params __a =flatten_dict(_snake_case ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __a =flatten_dict(flax_model.params['batch_stats'] ) random_flax_state_dict.update(_snake_case ) __a ={} __a =(model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __a =(model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __a =tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __a =pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __a =pt_tuple_key[1:] # Correctly rename weight parameters __a , __a =rename_key_and_reshape_tensor( _snake_case , _snake_case , _snake_case , _snake_case ) # add model prefix if necessary __a =(model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __a =(model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: __a =jnp.asarray(_snake_case ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_snake_case , _snake_case ) continue # also add unexpected weight so that warning is thrown __a =jnp.asarray(_snake_case ) else: # also add unexpected weight so that warning is thrown __a =jnp.asarray(_snake_case ) return unflatten_dict(_snake_case ) def UpperCamelCase_( _snake_case : int , _snake_case : str ): """simple docstring""" import torch # Load the index __a ={} for shard_file in shard_filenames: # load using msgpack utils __a =torch.load(_snake_case ) __a ={k: v.numpy() for k, v in pt_state_dict.items()} __a =flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __a =flax_model.params['params'] __a =flatten_dict(_snake_case ) random_flax_state_dict.update(flatten_dict(flax_model.params['batch_stats'] ) ) else: __a =flax_model.params __a =flatten_dict(_snake_case ) __a =(model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __a =(model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __a =tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __a =pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __a =pt_tuple_key[1:] # Correctly rename weight parameters __a , __a =rename_key_and_reshape_tensor( _snake_case , _snake_case , _snake_case , _snake_case ) # add model prefix if necessary __a =(model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __a =(model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: __a =jnp.asarray(_snake_case ) continue if "var" in flax_key[-1]: __a =jnp.asarray(_snake_case ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_snake_case , _snake_case ) continue # also add unexpected weight so that warning is thrown __a =jnp.asarray(_snake_case ) else: # also add unexpected weight so that warning is thrown __a =jnp.asarray(_snake_case ) return unflatten_dict(_snake_case ) def UpperCamelCase_( _snake_case : Union[str, Any] , _snake_case : Optional[Any] ): """simple docstring""" __a =os.path.abspath(_snake_case ) logger.info(F'Loading Flax weights from {flax_checkpoint_path}' ) # import correct flax class __a =getattr(_snake_case , 'Flax' + model.__class__.__name__ ) # load flax weight dict with open(_snake_case , 'rb' ) as state_f: try: __a =from_bytes(_snake_case , state_f.read() ) except UnpicklingError: raise EnvironmentError(F'Unable to convert {flax_checkpoint_path} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(_snake_case , _snake_case ) def UpperCamelCase_( _snake_case : Tuple , _snake_case : Dict ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights __a =flatten_dict(jax.tree_util.tree_map(lambda _snake_case : x.dtype == jnp.bfloataa , _snake_case ) ).values() if any(_snake_case ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) __a =jax.tree_util.tree_map( lambda _snake_case : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , _snake_case ) __a =flatten_dict(_snake_case ) __a =pt_model.state_dict() __a =(pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split('.' )[0] for k in pt_model_dict.keys()} ) __a =(pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split('.' )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys __a =[] __a =set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): __a =flax_key_tuple[0] == pt_model.base_model_prefix __a ='.'.join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: __a =flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: __a =(pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(_snake_case ) not in pt_model_dict: # conv layer __a =flax_key_tuple[:-1] + ('weight',) __a =jnp.transpose(_snake_case , (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(_snake_case ) not in pt_model_dict: # linear layer __a =flax_key_tuple[:-1] + ('weight',) __a =flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: __a =flax_key_tuple[:-1] + ('weight',) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: __a =flax_key_tuple[:-1] + ('running_mean',) elif "var" in flax_key_tuple[-1]: __a =flax_key_tuple[:-1] + ('running_var',) if "batch_stats" in flax_state: __a ='.'.join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: __a ='.'.join(_snake_case ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. __a ={} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: __a =key.split('.' ) __a =None if key_components[-3::2] == ["parametrizations", "original0"]: __a =key_components[-2] + '_g' elif key_components[-3::2] == ["parametrizations", "original1"]: __a =key_components[-2] + '_v' if name is not None: __a =key_components[:-3] + [name] __a ='.'.join(_snake_case ) __a =key if flax_key in special_pt_names: __a =special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' F'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict __a =np.asarray(_snake_case ) if not isinstance(_snake_case , np.ndarray ) else flax_tensor __a =torch.from_numpy(_snake_case ) # remove from missing keys missing_keys.remove(_snake_case ) else: # weight is not expected by PyTorch model unexpected_keys.append(_snake_case ) pt_model.load_state_dict(_snake_case ) # re-transform missing_keys to list __a =list(_snake_case ) if len(_snake_case ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' F' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) else: logger.warning(F'All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n' ) if len(_snake_case ) > 0: logger.warning( F'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' F' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ' use it for predictions and inference.' ) else: logger.warning( F'All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n' 'If your task is similar to the task the model of the checkpoint was trained on, ' F'you can already use {pt_model.__class__.__name__} for predictions without further training.' ) return pt_model
308
import os def UpperCamelCase_( _snake_case : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_snake_case ) , _snake_case ) ) as input_file: __a =[ [int(_snake_case ) for element in line.split(',' )] for line in input_file.readlines() ] __a =len(_snake_case ) __a =len(matrix[0] ) __a =[[-1 for _ in range(_snake_case )] for _ in range(_snake_case )] for i in range(_snake_case ): __a =matrix[i][0] for j in range(1 , _snake_case ): for i in range(_snake_case ): __a =minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , _snake_case ): __a =min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): __a =min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(f'''{solution() = }''')
308
1
'''simple docstring''' def UpperCamelCase_ ( _UpperCAmelCase : str , _UpperCAmelCase : list[str] ) -> str: """simple docstring""" _UpperCAmelCase : Optional[int] = "" for word_or_phrase in separated: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise Exception("join() accepts only strings to be joined" ) joined += word_or_phrase + separator return joined.strip(_UpperCAmelCase ) if __name__ == "__main__": from doctest import testmod testmod()
31
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class lowerCamelCase_ (unittest.TestCase ): '''simple docstring''' def __init__( self : int , A : Dict , A : Optional[int]=7 , A : Tuple=3 , A : Optional[Any]=10 , A : int=18 , A : Dict=30 , A : List[str]=400 , A : int=True , A : Optional[Any]=None , A : Optional[Any]=True , A : List[Any]=[0.5, 0.5, 0.5] , A : List[str]=[0.5, 0.5, 0.5] , A : Optional[int]=None , ): _UpperCAmelCase : Dict = size if size is not None else {"shortest_edge": 18} _UpperCAmelCase : Optional[Any] = crop_size if crop_size is not None else {"height": 18, "width": 18} _UpperCAmelCase : Tuple = parent _UpperCAmelCase : Any = batch_size _UpperCAmelCase : Optional[int] = num_channels _UpperCAmelCase : Optional[Any] = num_frames _UpperCAmelCase : Any = image_size _UpperCAmelCase : Dict = min_resolution _UpperCAmelCase : Any = max_resolution _UpperCAmelCase : Optional[int] = do_resize _UpperCAmelCase : str = size _UpperCAmelCase : List[Any] = do_normalize _UpperCAmelCase : Any = image_mean _UpperCAmelCase : Tuple = image_std _UpperCAmelCase : Any = crop_size def _A ( self : List[Any] ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class lowerCamelCase_ (snake_case__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase: Dict = VivitImageProcessor if is_vision_available() else None def _A ( self : int ): _UpperCAmelCase : Tuple = VivitImageProcessingTester(self ) @property def _A ( self : Optional[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def _A ( self : Union[str, Any] ): _UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , "image_mean" ) ) self.assertTrue(hasattr(A , "image_std" ) ) self.assertTrue(hasattr(A , "do_normalize" ) ) self.assertTrue(hasattr(A , "do_resize" ) ) self.assertTrue(hasattr(A , "do_center_crop" ) ) self.assertTrue(hasattr(A , "size" ) ) def _A ( self : List[Any] ): _UpperCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) _UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def _A ( self : Tuple ): # Initialize image_processing _UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos _UpperCAmelCase : Any = prepare_video_inputs(self.image_processor_tester , equal_resolution=A ) for video in video_inputs: self.assertIsInstance(A , A ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input _UpperCAmelCase : str = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _UpperCAmelCase : List[Any] = image_processing(A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def _A ( self : List[Any] ): # Initialize image_processing _UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCAmelCase : List[Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for video in video_inputs: self.assertIsInstance(A , A ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input _UpperCAmelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _UpperCAmelCase : Optional[int] = image_processing(A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def _A ( self : List[Any] ): # Initialize image_processing _UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCAmelCase : Optional[int] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for video in video_inputs: self.assertIsInstance(A , A ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input _UpperCAmelCase : Optional[Any] = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _UpperCAmelCase : List[Any] = image_processing(A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
31
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu snake_case_ : Optional[Any] = False class lowercase__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' return 12 @property def UpperCamelCase_ ( self : str ): '''simple docstring''' return 12 @property def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' return 32 @property def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : str = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,num_vq_embeddings=self.num_embed ,vq_embed_dim=3 ,) return model @property def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def UpperCamelCase_ ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : int = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=self.text_embedder_hidden_size ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_SCREAMING_SNAKE_CASE ) @property def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = 12 _UpperCamelCase : List[Any] = 12 _UpperCamelCase : int = { "attention_bias": True, "cross_attention_dim": 32, "attention_head_dim": height * width, "num_attention_heads": 1, "num_vector_embeds": self.num_embed, "num_embeds_ada_norm": self.num_embeds_ada_norm, "norm_num_groups": 32, "sample_size": width, "activation_fn": "geglu-approximate", } _UpperCamelCase : List[Any] = TransformeraDModel(**_SCREAMING_SNAKE_CASE ) return model def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Dict = "cpu" _UpperCamelCase : Optional[int] = self.dummy_vqvae _UpperCamelCase : str = self.dummy_text_encoder _UpperCamelCase : List[Any] = self.dummy_tokenizer _UpperCamelCase : Any = self.dummy_transformer _UpperCamelCase : Union[str, Any] = VQDiffusionScheduler(self.num_embed ) _UpperCamelCase : Any = LearnedClassifierFreeSamplingEmbeddings(learnable=_SCREAMING_SNAKE_CASE ) _UpperCamelCase : Optional[int] = VQDiffusionPipeline( vqvae=_SCREAMING_SNAKE_CASE ,text_encoder=_SCREAMING_SNAKE_CASE ,tokenizer=_SCREAMING_SNAKE_CASE ,transformer=_SCREAMING_SNAKE_CASE ,scheduler=_SCREAMING_SNAKE_CASE ,learned_classifier_free_sampling_embeddings=_SCREAMING_SNAKE_CASE ,) _UpperCamelCase : Optional[int] = pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _UpperCamelCase : str = "teddy bear playing in the pool" _UpperCamelCase : List[Any] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(0 ) _UpperCamelCase : List[Any] = pipe([prompt] ,generator=_SCREAMING_SNAKE_CASE ,num_inference_steps=2 ,output_type='np' ) _UpperCamelCase : List[Any] = output.images _UpperCamelCase : Optional[int] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(0 ) _UpperCamelCase : Tuple = pipe( [prompt] ,generator=_SCREAMING_SNAKE_CASE ,output_type='np' ,return_dict=_SCREAMING_SNAKE_CASE ,num_inference_steps=2 )[0] _UpperCamelCase : Any = image[0, -3:, -3:, -1] _UpperCamelCase : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) _UpperCamelCase : Tuple = np.array([0.6_5_5_1, 0.6_1_6_8, 0.5_0_0_8, 0.5_6_7_6, 0.5_6_5_9, 0.4_2_9_5, 0.6_0_7_3, 0.5_5_9_9, 0.4_9_9_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : int = "cpu" _UpperCamelCase : int = self.dummy_vqvae _UpperCamelCase : Union[str, Any] = self.dummy_text_encoder _UpperCamelCase : List[str] = self.dummy_tokenizer _UpperCamelCase : Optional[Any] = self.dummy_transformer _UpperCamelCase : List[str] = VQDiffusionScheduler(self.num_embed ) _UpperCamelCase : Optional[int] = LearnedClassifierFreeSamplingEmbeddings( learnable=_SCREAMING_SNAKE_CASE ,hidden_size=self.text_embedder_hidden_size ,length=tokenizer.model_max_length ) _UpperCamelCase : Union[str, Any] = VQDiffusionPipeline( vqvae=_SCREAMING_SNAKE_CASE ,text_encoder=_SCREAMING_SNAKE_CASE ,tokenizer=_SCREAMING_SNAKE_CASE ,transformer=_SCREAMING_SNAKE_CASE ,scheduler=_SCREAMING_SNAKE_CASE ,learned_classifier_free_sampling_embeddings=_SCREAMING_SNAKE_CASE ,) _UpperCamelCase : str = pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _UpperCamelCase : Optional[Any] = "teddy bear playing in the pool" _UpperCamelCase : Optional[int] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(0 ) _UpperCamelCase : Optional[int] = pipe([prompt] ,generator=_SCREAMING_SNAKE_CASE ,num_inference_steps=2 ,output_type='np' ) _UpperCamelCase : str = output.images _UpperCamelCase : Optional[Any] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(0 ) _UpperCamelCase : Any = pipe( [prompt] ,generator=_SCREAMING_SNAKE_CASE ,output_type='np' ,return_dict=_SCREAMING_SNAKE_CASE ,num_inference_steps=2 )[0] _UpperCamelCase : str = image[0, -3:, -3:, -1] _UpperCamelCase : List[str] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) _UpperCamelCase : Any = np.array([0.6_6_9_3, 0.6_0_7_5, 0.4_9_5_9, 0.5_7_0_1, 0.5_5_8_3, 0.4_3_3_3, 0.6_1_7_1, 0.5_6_8_4, 0.4_9_8_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase_ ( self : Dict ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy' ) _UpperCamelCase : Optional[Any] = VQDiffusionPipeline.from_pretrained('microsoft/vq-diffusion-ithq' ) _UpperCamelCase : Optional[Any] = pipeline.to(_SCREAMING_SNAKE_CASE ) pipeline.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though _UpperCamelCase : Optional[int] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipeline( 'teddy bear playing in the pool' ,num_images_per_prompt=1 ,generator=_SCREAMING_SNAKE_CASE ,output_type='np' ,) _UpperCamelCase : str = output.images[0] assert image.shape == (256, 256, 3) assert np.abs(expected_image - image ).max() < 2.0
364
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case_ : int = logging.get_logger(__name__) def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = DPTConfig(embedding_type='hybrid' ) if "large" in checkpoint_url: _UpperCamelCase : Any = 1_0_2_4 _UpperCamelCase : List[Any] = 4_0_9_6 _UpperCamelCase : List[str] = 2_4 _UpperCamelCase : Tuple = 1_6 _UpperCamelCase : Union[str, Any] = [5, 1_1, 1_7, 2_3] _UpperCamelCase : Any = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] _UpperCamelCase : Tuple = (1, 3_8_4, 3_8_4) if "nyu" or "midas" in checkpoint_url: _UpperCamelCase : Optional[int] = 7_6_8 _UpperCamelCase : Optional[Any] = [1, 1, 1, 0.5] _UpperCamelCase : List[Any] = [2_5_6, 5_1_2, 7_6_8, 7_6_8] _UpperCamelCase : Optional[int] = 1_5_0 _UpperCamelCase : Tuple = 1_6 _UpperCamelCase : Dict = (1, 3_8_4, 3_8_4) _UpperCamelCase : Optional[int] = False _UpperCamelCase : Optional[int] = 'project' if "ade" in checkpoint_url: _UpperCamelCase : Dict = True _UpperCamelCase : Dict = 7_6_8 _UpperCamelCase : Union[str, Any] = [1, 1, 1, 0.5] _UpperCamelCase : Union[str, Any] = 1_5_0 _UpperCamelCase : str = 1_6 _UpperCamelCase : Tuple = 'huggingface/label-files' _UpperCamelCase : Tuple = 'ade20k-id2label.json' _UpperCamelCase : Tuple = json.load(open(cached_download(hf_hub_url(UpperCAmelCase_ , UpperCAmelCase_ , repo_type='dataset' ) ) , 'r' ) ) _UpperCamelCase : str = {int(UpperCAmelCase_ ): v for k, v in idalabel.items()} _UpperCamelCase : List[str] = idalabel _UpperCamelCase : Union[str, Any] = {v: k for k, v in idalabel.items()} _UpperCamelCase : int = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = ['pretrained.model.head.weight', 'pretrained.model.head.bias'] for k in ignore_keys: state_dict.pop(UpperCAmelCase_ , UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): _UpperCamelCase : List[str] = name.replace('pretrained.model' , 'dpt.encoder' ) if "pretrained.model" in name: _UpperCamelCase : int = name.replace('pretrained.model' , 'dpt.embeddings' ) if "patch_embed" in name: _UpperCamelCase : Any = name.replace('patch_embed' , '' ) if "pos_embed" in name: _UpperCamelCase : Tuple = name.replace('pos_embed' , 'position_embeddings' ) if "attn.proj" in name: _UpperCamelCase : List[str] = name.replace('attn.proj' , 'attention.output.dense' ) if "proj" in name and "project" not in name: _UpperCamelCase : int = name.replace('proj' , 'projection' ) if "blocks" in name: _UpperCamelCase : List[str] = name.replace('blocks' , 'layer' ) if "mlp.fc1" in name: _UpperCamelCase : Union[str, Any] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _UpperCamelCase : str = name.replace('mlp.fc2' , 'output.dense' ) if "norm1" in name and "backbone" not in name: _UpperCamelCase : Tuple = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name and "backbone" not in name: _UpperCamelCase : str = name.replace('norm2' , 'layernorm_after' ) if "scratch.output_conv" in name: _UpperCamelCase : Dict = name.replace('scratch.output_conv' , 'head' ) if "scratch" in name: _UpperCamelCase : List[str] = name.replace('scratch' , 'neck' ) if "layer1_rn" in name: _UpperCamelCase : List[str] = name.replace('layer1_rn' , 'convs.0' ) if "layer2_rn" in name: _UpperCamelCase : Any = name.replace('layer2_rn' , 'convs.1' ) if "layer3_rn" in name: _UpperCamelCase : int = name.replace('layer3_rn' , 'convs.2' ) if "layer4_rn" in name: _UpperCamelCase : Dict = name.replace('layer4_rn' , 'convs.3' ) if "refinenet" in name: _UpperCamelCase : str = int(name[len('neck.refinenet' ) : len('neck.refinenet' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 _UpperCamelCase : str = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: _UpperCamelCase : Dict = name.replace('out_conv' , 'projection' ) if "resConfUnit1" in name: _UpperCamelCase : Union[str, Any] = name.replace('resConfUnit1' , 'residual_layer1' ) if "resConfUnit2" in name: _UpperCamelCase : Union[str, Any] = name.replace('resConfUnit2' , 'residual_layer2' ) if "conv1" in name: _UpperCamelCase : int = name.replace('conv1' , 'convolution1' ) if "conv2" in name: _UpperCamelCase : Dict = name.replace('conv2' , 'convolution2' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: _UpperCamelCase : str = name.replace('pretrained.act_postprocess1.0.project.0' , 'neck.reassemble_stage.readout_projects.0.0' ) if "pretrained.act_postprocess2.0.project.0" in name: _UpperCamelCase : Union[str, Any] = name.replace('pretrained.act_postprocess2.0.project.0' , 'neck.reassemble_stage.readout_projects.1.0' ) if "pretrained.act_postprocess3.0.project.0" in name: _UpperCamelCase : Optional[int] = name.replace('pretrained.act_postprocess3.0.project.0' , 'neck.reassemble_stage.readout_projects.2.0' ) if "pretrained.act_postprocess4.0.project.0" in name: _UpperCamelCase : Optional[int] = name.replace('pretrained.act_postprocess4.0.project.0' , 'neck.reassemble_stage.readout_projects.3.0' ) # resize blocks if "pretrained.act_postprocess1.3" in name: _UpperCamelCase : Union[str, Any] = name.replace('pretrained.act_postprocess1.3' , 'neck.reassemble_stage.layers.0.projection' ) if "pretrained.act_postprocess1.4" in name: _UpperCamelCase : List[str] = name.replace('pretrained.act_postprocess1.4' , 'neck.reassemble_stage.layers.0.resize' ) if "pretrained.act_postprocess2.3" in name: _UpperCamelCase : List[Any] = name.replace('pretrained.act_postprocess2.3' , 'neck.reassemble_stage.layers.1.projection' ) if "pretrained.act_postprocess2.4" in name: _UpperCamelCase : List[Any] = name.replace('pretrained.act_postprocess2.4' , 'neck.reassemble_stage.layers.1.resize' ) if "pretrained.act_postprocess3.3" in name: _UpperCamelCase : Dict = name.replace('pretrained.act_postprocess3.3' , 'neck.reassemble_stage.layers.2.projection' ) if "pretrained.act_postprocess4.3" in name: _UpperCamelCase : Union[str, Any] = name.replace('pretrained.act_postprocess4.3' , 'neck.reassemble_stage.layers.3.projection' ) if "pretrained.act_postprocess4.4" in name: _UpperCamelCase : List[str] = name.replace('pretrained.act_postprocess4.4' , 'neck.reassemble_stage.layers.3.resize' ) if "pretrained" in name: _UpperCamelCase : int = name.replace('pretrained' , 'dpt' ) if "bn" in name: _UpperCamelCase : Union[str, Any] = name.replace('bn' , 'batch_norm' ) if "head" in name: _UpperCamelCase : Dict = name.replace('head' , 'head.head' ) if "encoder.norm" in name: _UpperCamelCase : str = name.replace('encoder.norm' , 'layernorm' ) if "auxlayer" in name: _UpperCamelCase : Any = name.replace('auxlayer' , 'auxiliary_head.head' ) if "backbone" in name: _UpperCamelCase : List[Any] = name.replace('backbone' , 'backbone.bit.encoder' ) if ".." in name: _UpperCamelCase : Dict = name.replace('..' , '.' ) if "stem.conv" in name: _UpperCamelCase : Tuple = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: _UpperCamelCase : Optional[int] = name.replace('blocks' , 'layers' ) if "convolution" in name and "backbone" in name: _UpperCamelCase : List[str] = name.replace('convolution' , 'conv' ) if "layer" in name and "backbone" in name: _UpperCamelCase : Union[str, Any] = name.replace('layer' , 'layers' ) if "backbone.bit.encoder.bit" in name: _UpperCamelCase : Dict = name.replace('backbone.bit.encoder.bit' , 'backbone.bit' ) if "embedder.conv" in name: _UpperCamelCase : str = name.replace('embedder.conv' , 'embedder.convolution' ) if "backbone.bit.encoder.stem.norm" in name: _UpperCamelCase : Tuple = name.replace('backbone.bit.encoder.stem.norm' , 'backbone.bit.embedder.norm' ) return name def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase : List[str] = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) _UpperCamelCase : List[str] = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase : List[str] = in_proj_weight[: config.hidden_size, :] _UpperCamelCase : int = in_proj_bias[: config.hidden_size] _UpperCamelCase : str = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase : Union[str, Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase : List[Any] = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def A__ ( ): _UpperCamelCase : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCamelCase : List[Any] = Image.open(requests.get(UpperCAmelCase_ , stream=UpperCAmelCase_ ).raw ) return im @torch.no_grad() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : int = get_dpt_config(UpperCAmelCase_ ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") _UpperCamelCase : List[str] = torch.load(UpperCAmelCase_ , map_location='cpu' ) # remove certain keys remove_ignore_keys_(UpperCAmelCase_ ) # rename keys for key in state_dict.copy().keys(): _UpperCamelCase : Any = state_dict.pop(UpperCAmelCase_ ) _UpperCamelCase : int = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase_ , UpperCAmelCase_ ) # load HuggingFace model _UpperCamelCase : Union[str, Any] = DPTForSemanticSegmentation(UpperCAmelCase_ ) if 'ade' in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase_ ) model.load_state_dict(UpperCAmelCase_ ) model.eval() # Check outputs on an image _UpperCamelCase : Tuple = 4_8_0 if 'ade' in checkpoint_url else 3_8_4 _UpperCamelCase : Any = DPTImageProcessor(size=UpperCAmelCase_ ) _UpperCamelCase : Union[str, Any] = prepare_img() _UpperCamelCase : Optional[int] = image_processor(UpperCAmelCase_ , return_tensors='pt' ) # forward pass _UpperCamelCase : Optional[Any] = model(**UpperCAmelCase_ ).logits if 'ade' in checkpoint_url else model(**UpperCAmelCase_ ).predicted_depth if show_prediction: _UpperCamelCase : List[str] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='bicubic' , align_corners=UpperCAmelCase_ , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 2_5_5 ).show() if pytorch_dump_folder_path is not None: Path(UpperCAmelCase_ ).mkdir(exist_ok=UpperCAmelCase_ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase_ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase_ ) if push_to_hub: model.push_to_hub('ybelkada/dpt-hybrid-midas' ) image_processor.push_to_hub('ybelkada/dpt-hybrid-midas' ) if __name__ == "__main__": snake_case_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=False, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) parser.add_argument( '--show_prediction', action='store_true', ) snake_case_ : Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
236
0
"""simple docstring""" from __future__ import annotations class lowerCAmelCase__ : '''simple docstring''' def __init__( self : Any , lowercase_ : int = 0): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = key def _SCREAMING_SNAKE_CASE ( self : Any , lowercase_ : str , lowercase_ : int): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) SCREAMING_SNAKE_CASE_ : Dict = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(lowercase_) ^ key) for ch in content] def _SCREAMING_SNAKE_CASE ( self : str , lowercase_ : str , lowercase_ : int): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) SCREAMING_SNAKE_CASE_ : Union[str, Any] = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(lowercase_) ^ key) for ch in content] def _SCREAMING_SNAKE_CASE ( self : Tuple , lowercase_ : str , lowercase_ : int = 0): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) SCREAMING_SNAKE_CASE_ : int = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned SCREAMING_SNAKE_CASE_ : List[str] = '''''' for ch in content: ans += chr(ord(lowercase_) ^ key) return ans def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : str , lowercase_ : int = 0): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) SCREAMING_SNAKE_CASE_ : Union[str, Any] = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned SCREAMING_SNAKE_CASE_ : List[Any] = '''''' for ch in content: ans += chr(ord(lowercase_) ^ key) return ans def _SCREAMING_SNAKE_CASE ( self : str , lowercase_ : str , lowercase_ : int = 0): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) try: with open(lowercase_) as fin, open('''encrypt.out''' , '''w+''') as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(lowercase_ , lowercase_)) except OSError: return False return True def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowercase_ : str , lowercase_ : int): '''simple docstring''' assert isinstance(lowercase_ , lowercase_) and isinstance(lowercase_ , lowercase_) try: with open(lowercase_) as fin, open('''decrypt.out''' , '''w+''') as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(lowercase_ , lowercase_)) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
91
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run a_ = True except (ImportError, AttributeError): a_ = object def __UpperCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): pass a_ = False a_ = logging.get_logger('transformers-cli/serving') def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : Optional[Any] = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(__UpperCamelCase , args.host , args.port , args.workers ) class UpperCAmelCase_ ( snake_case ): UpperCamelCase =42 class UpperCAmelCase_ ( snake_case ): UpperCamelCase =42 UpperCamelCase =42 class UpperCAmelCase_ ( snake_case ): UpperCamelCase =42 class UpperCAmelCase_ ( snake_case ): UpperCamelCase =42 class UpperCAmelCase_ ( snake_case ): @staticmethod def _lowerCamelCase ( UpperCamelCase_ ) -> Tuple: __lowercase : Dict = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=UpperCamelCase_ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=UpperCamelCase_ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=UpperCamelCase_ , default=88_88 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=UpperCamelCase_ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=UpperCamelCase_ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=UpperCamelCase_ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=UpperCamelCase_ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=UpperCamelCase_ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=UpperCamelCase_ ) def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Any: __lowercase : List[Any] = pipeline __lowercase : str = host __lowercase : List[str] = port __lowercase : str = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) __lowercase : int = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=UpperCamelCase_ , response_class=UpperCamelCase_ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=UpperCamelCase_ , response_class=UpperCamelCase_ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=UpperCamelCase_ , response_class=UpperCamelCase_ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=UpperCamelCase_ , response_class=UpperCamelCase_ , methods=['''POST'''] , ), ] , timeout=6_00 , ) def _lowerCamelCase ( self ) -> Union[str, Any]: run(self._app , host=self.host , port=self.port , workers=self.workers ) def _lowerCamelCase ( self ) -> Tuple: return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def _lowerCamelCase ( self , UpperCamelCase_ = Body(UpperCamelCase_ , embed=UpperCamelCase_ ) , UpperCamelCase_ = Body(UpperCamelCase_ , embed=UpperCamelCase_ ) ) -> Optional[int]: try: __lowercase : Any = self._pipeline.tokenizer.tokenize(UpperCamelCase_ ) if return_ids: __lowercase : Dict = self._pipeline.tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) return ServeTokenizeResult(tokens=UpperCamelCase_ , tokens_ids=UpperCamelCase_ ) else: return ServeTokenizeResult(tokens=UpperCamelCase_ ) except Exception as e: raise HTTPException(status_code=5_00 , detail={'''model''': '''''', '''error''': str(UpperCamelCase_ )} ) def _lowerCamelCase ( self , UpperCamelCase_ = Body(UpperCamelCase_ , embed=UpperCamelCase_ ) , UpperCamelCase_ = Body(UpperCamelCase_ , embed=UpperCamelCase_ ) , UpperCamelCase_ = Body(UpperCamelCase_ , embed=UpperCamelCase_ ) , ) -> Dict: try: __lowercase : Tuple = self._pipeline.tokenizer.decode(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) return ServeDeTokenizeResult(model='''''' , text=UpperCamelCase_ ) except Exception as e: raise HTTPException(status_code=5_00 , detail={'''model''': '''''', '''error''': str(UpperCamelCase_ )} ) async def _lowerCamelCase ( self , UpperCamelCase_=Body(UpperCamelCase_ , embed=UpperCamelCase_ ) ) -> Union[str, Any]: # Check we don't have empty string if len(UpperCamelCase_ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model __lowercase : Optional[Any] = self._pipeline(UpperCamelCase_ ) return ServeForwardResult(output=UpperCamelCase_ ) except Exception as e: raise HTTPException(5_00 , {'''error''': str(UpperCamelCase_ )} )
249
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class __a ( __UpperCamelCase ): __snake_case : int = """openai/whisper-base""" __snake_case : Optional[int] = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) __snake_case : List[str] = """transcriber""" __snake_case : str = WhisperProcessor __snake_case : Optional[Any] = WhisperForConditionalGeneration __snake_case : List[Any] = ["""audio"""] __snake_case : List[str] = ["""text"""] def A ( self : int , UpperCAmelCase : List[str] ): return self.pre_processor(UpperCAmelCase , return_tensors="""pt""" ).input_features def A ( self : Union[str, Any] , UpperCAmelCase : Tuple ): return self.model.generate(inputs=UpperCAmelCase ) def A ( self : Optional[int] , UpperCAmelCase : List[str] ): return self.pre_processor.batch_decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase )[0]
363
from math import ceil def __UpperCamelCase ( lowercase__ : int = 1001 ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): lowerCAmelCase_ : Optional[Any] = 2 * i + 1 lowerCAmelCase_ : Union[str, Any] = 2 * i lowerCAmelCase_ : Optional[Any] = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: __UpperCAmelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('Invalid entry - please enter a number')
28
0
'''simple docstring''' from __future__ import annotations import os from collections.abc import Mapping __a: Tuple = tuple[int, int] class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: lowercase__ : str = vertices lowercase__ : Optional[int] = { (min(__UpperCAmelCase ), max(__UpperCAmelCase )): weight for edge, weight in edges.items() } def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> int: self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) lowercase__ : Tuple = weight def _lowerCAmelCase( self ) -> Optional[int]: lowercase__ : List[str] = Graph({min(self.vertices )} , {} ) lowercase__ : str = 42 lowercase__ : Optional[int] = 42 lowercase__ : Optional[int] = 42 lowercase__ : Optional[int] = 42 while len(subgraph.vertices ) < len(self.vertices ): lowercase__ : List[str] = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: lowercase__ : Union[str, Any] = edge lowercase__ : List[str] = weight subgraph.add_edge(__UpperCAmelCase , __UpperCAmelCase ) return subgraph def __UpperCamelCase ( UpperCAmelCase = "p107_network.txt" ): lowercase__ : Tuple = os.path.abspath(os.path.dirname(_UpperCamelCase ) ) lowercase__ : Dict = os.path.join(_UpperCamelCase , _UpperCamelCase ) lowercase__ : Dict = {} lowercase__ : Any = 42 lowercase__ : int = 42 lowercase__ : List[Any] = 42 with open(_UpperCamelCase ) as f: lowercase__ : str = f.read().strip().split('''\n''' ) lowercase__ : Dict = [line.split(''',''' ) for line in data] for edgea in range(1 , len(_UpperCamelCase ) ): for edgea in range(_UpperCamelCase ): if adjaceny_matrix[edgea][edgea] != "-": lowercase__ : Union[str, Any] = int(adjaceny_matrix[edgea][edgea] ) lowercase__ : Dict = Graph(set(range(len(_UpperCamelCase ) ) ) , _UpperCamelCase ) lowercase__ : Optional[Any] = graph.prims_algorithm() lowercase__ : Union[str, Any] = sum(graph.edges.values() ) lowercase__ : Tuple = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F'{solution() = }')
198
from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING a_ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase__ ) class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' super().__init__(*__UpperCAmelCase , **__UpperCAmelCase ) requires_backends(self , '''vision''' ) self.check_model_type(__UpperCAmelCase ) def __call__( self , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return super().__call__(__UpperCAmelCase , **__UpperCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return {}, {}, {} def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = load_image(__UpperCAmelCase ) __lowerCamelCase = image.size __lowerCamelCase = self.image_processor(images=__UpperCAmelCase , return_tensors=self.framework ) return model_inputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.model(**__UpperCAmelCase ) return model_outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = model_outputs.predicted_depth __lowerCamelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode='''bicubic''' , align_corners=__UpperCAmelCase ) __lowerCamelCase = prediction.squeeze().cpu().numpy() __lowerCamelCase = (output * 255 / np.max(__UpperCAmelCase )).astype('''uint8''' ) __lowerCamelCase = Image.fromarray(__UpperCAmelCase ) __lowerCamelCase = {} __lowerCamelCase = predicted_depth __lowerCamelCase = depth return output_dict
330
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _SCREAMING_SNAKE_CASE ( __a ,unittest.TestCase ): __SCREAMING_SNAKE_CASE :List[str] = KandinskyVaaPipeline __SCREAMING_SNAKE_CASE :Optional[Any] = [ """image_embeds""", """negative_image_embeds""", ] __SCREAMING_SNAKE_CASE :Tuple = ["""image_embeds""", """negative_image_embeds"""] __SCREAMING_SNAKE_CASE :Tuple = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __SCREAMING_SNAKE_CASE :List[str] = False @property def snake_case__ ( self : List[Any] ): return 32 @property def snake_case__ ( self : Dict ): return 32 @property def snake_case__ ( self : Optional[Any] ): return self.time_input_dim @property def snake_case__ ( self : Optional[Any] ): return self.time_input_dim * 4 @property def snake_case__ ( self : Dict ): return 100 @property def snake_case__ ( self : Tuple ): torch.manual_seed(0 ) __magic_name__ = { '''in_channels''': 4, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } __magic_name__ = UNetaDConditionModel(**a__ ) return model @property def snake_case__ ( self : Any ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def snake_case__ ( self : str ): torch.manual_seed(0 ) __magic_name__ = VQModel(**self.dummy_movq_kwargs ) return model def snake_case__ ( self : Any ): __magic_name__ = self.dummy_unet __magic_name__ = self.dummy_movq __magic_name__ = DDIMScheduler( num_train_timesteps=1000 , beta_schedule='''linear''' , beta_start=0.00_085 , beta_end=0.012 , clip_sample=a__ , set_alpha_to_one=a__ , steps_offset=1 , prediction_type='''epsilon''' , thresholding=a__ , ) __magic_name__ = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def snake_case__ ( self : Any , a__ : Union[str, Any] , a__ : Any=0 ): __magic_name__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(a__ ) ).to(a__ ) __magic_name__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( a__ ) if str(a__ ).startswith('''mps''' ): __magic_name__ = torch.manual_seed(a__ ) else: __magic_name__ = torch.Generator(device=a__ ).manual_seed(a__ ) __magic_name__ = { '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''generator''': generator, '''height''': 64, '''width''': 64, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def snake_case__ ( self : List[str] ): __magic_name__ = '''cpu''' __magic_name__ = self.get_dummy_components() __magic_name__ = self.pipeline_class(**a__ ) __magic_name__ = pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) __magic_name__ = pipe(**self.get_dummy_inputs(a__ ) ) __magic_name__ = output.images __magic_name__ = pipe( **self.get_dummy_inputs(a__ ) , return_dict=a__ , )[0] __magic_name__ = image[0, -3:, -3:, -1] __magic_name__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __magic_name__ = np.array( [0.6_237_976, 1.0, 0.36_441_332, 1.0, 0.70_639_634, 0.29_877_186, 0.85_652_125, 0.5_216_843, 0.54_454_046] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def snake_case__ ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : Dict ): __magic_name__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy''' ) __magic_name__ = KandinskyVaaPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(a__ ) __magic_name__ = KandinskyVaaPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-decoder''' , torch_dtype=torch.floataa ) __magic_name__ = pipeline.to(a__ ) pipeline.set_progress_bar_config(disable=a__ ) __magic_name__ = '''red cat, 4k photo''' __magic_name__ = torch.Generator(device='''cuda''' ).manual_seed(0 ) __magic_name__ , __magic_name__ = pipe_prior( a__ , generator=a__ , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() __magic_name__ = torch.Generator(device='''cuda''' ).manual_seed(0 ) __magic_name__ = pipeline( image_embeds=a__ , negative_image_embeds=a__ , generator=a__ , num_inference_steps=100 , output_type='''np''' , ) __magic_name__ = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(a__ , a__ )
98
'''simple docstring''' import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def UpperCamelCase ( a , a , a , a=1024 ) -> Union[str, Any]: '''simple docstring''' __magic_name__ , __magic_name__ = [], [] __magic_name__ = list(zip(a , a ) ) __magic_name__ , __magic_name__ = sorted_examples[0] def is_too_big(a ): return tok(a , return_tensors='''pt''' ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): __magic_name__ = new_src + ''' ''' + src __magic_name__ = new_tgt + ''' ''' + tgt if is_too_big(a ) or is_too_big(a ): # cant fit, finalize example finished_src.append(a ) finished_tgt.append(a ) __magic_name__ , __magic_name__ = src, tgt else: # can fit, keep adding __magic_name__ , __magic_name__ = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(a ) finished_tgt.append(a ) return finished_src, finished_tgt def UpperCamelCase ( a , a , a , a ) -> Any: '''simple docstring''' __magic_name__ = Path(a ) save_path.mkdir(exist_ok=a ) for split in ["train"]: __magic_name__ , __magic_name__ = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' __magic_name__ = [x.rstrip() for x in Path(a ).open().readlines()] __magic_name__ = [x.rstrip() for x in Path(a ).open().readlines()] __magic_name__ , __magic_name__ = pack_examples(a , a , a , a ) print(F'''packed {split} split from {len(a )} examples -> {len(a )}.''' ) Path(save_path / F'''{split}.source''' ).open('''w''' ).write('''\n'''.join(a ) ) Path(save_path / F'''{split}.target''' ).open('''w''' ).write('''\n'''.join(a ) ) for split in ["val", "test"]: __magic_name__ , __magic_name__ = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' shutil.copyfile(a , save_path / F'''{split}.source''' ) shutil.copyfile(a , save_path / F'''{split}.target''' ) def UpperCamelCase ( ) -> List[str]: '''simple docstring''' __magic_name__ = argparse.ArgumentParser() parser.add_argument('''--tok_name''' , type=a , help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''--max_seq_len''' , type=a , default=128 ) parser.add_argument('''--data_dir''' , type=a ) parser.add_argument('''--save_path''' , type=a ) __magic_name__ = parser.parse_args() __magic_name__ = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(a , Path(args.data_dir ) , args.max_seq_len , args.save_path ) if __name__ == "__main__": packer_cli()
98
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' lowerCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): lowerCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): lowerCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] lowerCAmelCase = key.replace(F'patch_embed{idx}' , F'patch_embeddings.{int(SCREAMING_SNAKE_CASE )-1}' ) if "norm" in key: lowerCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] lowerCAmelCase = key.replace(F'layer_norm{idx}' , F'layer_norm.{int(SCREAMING_SNAKE_CASE )-1}' ) if "layer_norm1" in key: lowerCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: lowerCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 lowerCAmelCase = key[key.find("""block""" ) + len("""block""" )] lowerCAmelCase = key.replace(F'block{idx}' , F'block.{int(SCREAMING_SNAKE_CASE )-1}' ) if "attn.q" in key: lowerCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: lowerCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: lowerCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: lowerCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: lowerCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: lowerCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: lowerCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) lowerCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] lowerCAmelCase = key.replace(F'linear_c{idx}' , F'linear_c.{int(SCREAMING_SNAKE_CASE )-1}' ) if "bot_conv" in key: lowerCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: lowerCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: lowerCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: lowerCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: lowerCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: lowerCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: lowerCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): lowerCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) lowerCAmelCase = value return new_state_dict def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCAmelCase = state_dict.pop(F'glpn.encoder.block.{i}.{j}.attention.self.kv.weight' ) lowerCAmelCase = state_dict.pop(F'glpn.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict lowerCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] lowerCAmelCase = kv_bias[: config.hidden_sizes[i]] lowerCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] lowerCAmelCase = kv_bias[config.hidden_sizes[i] :] def UpperCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCAmelCase = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ) return image @torch.no_grad() def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, Any]=False , SCREAMING_SNAKE_CASE : Union[str, Any]=None ): '''simple docstring''' lowerCAmelCase = GLPNConfig(hidden_sizes=[64, 1_28, 3_20, 5_12] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCAmelCase = GLPNImageProcessor() # prepare image lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict lowerCAmelCase = torch.load(SCREAMING_SNAKE_CASE , map_location=torch.device("""cpu""" ) ) # rename keys lowerCAmelCase = rename_keys(SCREAMING_SNAKE_CASE ) # key and value matrices need special treatment read_in_k_v(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # create HuggingFace model and load state dict lowerCAmelCase = GLPNForDepthEstimation(SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) model.eval() # forward pass lowerCAmelCase = model(SCREAMING_SNAKE_CASE ) lowerCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCAmelCase = torch.tensor( [[4.41_47, 4.08_73, 4.06_73], [3.78_90, 3.28_81, 3.15_25], [3.76_74, 3.54_23, 3.49_13]] ) elif "kitti" in model_name: lowerCAmelCase = torch.tensor( [[3.42_91, 2.78_65, 2.51_51], [3.28_41, 2.70_21, 2.35_02], [3.11_47, 2.46_25, 2.24_81]] ) else: raise ValueError(F'Unknown model name: {model_name}' ) lowerCAmelCase = torch.Size([1, 4_80, 6_40] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=SCREAMING_SNAKE_CASE , ) image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=SCREAMING_SNAKE_CASE , ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) parser.add_argument( "--model_name", default="glpn-kitti", type=str, help="Name of the model in case you're pushing to the hub.", ) SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
46
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available SCREAMING_SNAKE_CASE__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
46
1
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { """facebook/detr-resnet-50""": """https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json""", # See all DETR models at https://huggingface.co/models?filter=detr } class a_ ( a_ ): '''simple docstring''' __a: List[Any] = '''detr''' __a: Optional[Any] = ['''past_key_values'''] __a: Tuple = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=1_0_0 , lowercase_=6 , lowercase_=2_0_4_8 , lowercase_=8 , lowercase_=6 , lowercase_=2_0_4_8 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=2_5_6 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ) -> Union[str, Any]: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) lowerCAmelCase_ = CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(lowercase_ , lowercase_ ): lowerCAmelCase_ = backbone_config.get('model_type' ) lowerCAmelCase_ = CONFIG_MAPPING[backbone_model_type] lowerCAmelCase_ = config_class.from_dict(lowercase_ ) # set timm attributes to None lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None, None, None lowerCAmelCase_ = use_timm_backbone lowerCAmelCase_ = backbone_config lowerCAmelCase_ = num_channels lowerCAmelCase_ = num_queries lowerCAmelCase_ = d_model lowerCAmelCase_ = encoder_ffn_dim lowerCAmelCase_ = encoder_layers lowerCAmelCase_ = encoder_attention_heads lowerCAmelCase_ = decoder_ffn_dim lowerCAmelCase_ = decoder_layers lowerCAmelCase_ = decoder_attention_heads lowerCAmelCase_ = dropout lowerCAmelCase_ = attention_dropout lowerCAmelCase_ = activation_dropout lowerCAmelCase_ = activation_function lowerCAmelCase_ = init_std lowerCAmelCase_ = init_xavier_std lowerCAmelCase_ = encoder_layerdrop lowerCAmelCase_ = decoder_layerdrop lowerCAmelCase_ = encoder_layers lowerCAmelCase_ = auxiliary_loss lowerCAmelCase_ = position_embedding_type lowerCAmelCase_ = backbone lowerCAmelCase_ = use_pretrained_backbone lowerCAmelCase_ = dilation # Hungarian matcher lowerCAmelCase_ = class_cost lowerCAmelCase_ = bbox_cost lowerCAmelCase_ = giou_cost # Loss coefficients lowerCAmelCase_ = mask_loss_coefficient lowerCAmelCase_ = dice_loss_coefficient lowerCAmelCase_ = bbox_loss_coefficient lowerCAmelCase_ = giou_loss_coefficient lowerCAmelCase_ = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def _lowercase ( self ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def _lowercase ( self ) -> int: '''simple docstring''' return self.d_model @classmethod def _lowercase ( cls , lowercase_ , **lowercase_ ) -> List[str]: '''simple docstring''' return cls(backbone_config=lowercase_ , **lowercase_ ) def _lowercase ( self ) -> Dict[str, any]: '''simple docstring''' lowerCAmelCase_ = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: lowerCAmelCase_ = self.backbone_config.to_dict() lowerCAmelCase_ = self.__class__.model_type return output class a_ ( a_ ): '''simple docstring''' __a: List[Any] = version.parse('''1.11''' ) @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def _lowercase ( self ) -> float: '''simple docstring''' return 1e-5 @property def _lowercase ( self ) -> int: '''simple docstring''' return 1_2
358
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { """shi-labs/nat-mini-in1k-224""": """https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json""", # See all Nat models at https://huggingface.co/models?filter=nat } class a_ ( a_ , a_ ): '''simple docstring''' __a: Optional[Any] = '''nat''' __a: int = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self , lowercase_=4 , lowercase_=3 , lowercase_=6_4 , lowercase_=[3, 4, 6, 5] , lowercase_=[2, 4, 8, 1_6] , lowercase_=7 , lowercase_=3.0 , lowercase_=True , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.1 , lowercase_="gelu" , lowercase_=0.02 , lowercase_=1e-5 , lowercase_=0.0 , lowercase_=None , lowercase_=None , **lowercase_ , ) -> Optional[int]: '''simple docstring''' super().__init__(**lowercase_ ) lowerCAmelCase_ = patch_size lowerCAmelCase_ = num_channels lowerCAmelCase_ = embed_dim lowerCAmelCase_ = depths lowerCAmelCase_ = len(lowercase_ ) lowerCAmelCase_ = num_heads lowerCAmelCase_ = kernel_size lowerCAmelCase_ = mlp_ratio lowerCAmelCase_ = qkv_bias lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = drop_path_rate lowerCAmelCase_ = hidden_act lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ = int(embed_dim * 2 ** (len(lowercase_ ) - 1) ) lowerCAmelCase_ = layer_scale_init_value lowerCAmelCase_ = ['stem'] + [f'''stage{idx}''' for idx in range(1 , len(lowercase_ ) + 1 )] lowerCAmelCase_ , lowerCAmelCase_ = get_aligned_output_features_output_indices( out_features=lowercase_ , out_indices=lowercase_ , stage_names=self.stage_names )
14
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __snake_case =TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __snake_case =TaTokenizerFast __snake_case ={"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =[ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __snake_case =_LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
4
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = {} class lowercase_ ( lowercase ): '''simple docstring''' __snake_case = '''llama''' __snake_case = ['''past_key_values'''] def __init__( self : Optional[Any] , __UpperCAmelCase : Union[str, Any]=32_000 , __UpperCAmelCase : str=4_096 , __UpperCAmelCase : int=11_008 , __UpperCAmelCase : Tuple=32 , __UpperCAmelCase : Optional[int]=32 , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Union[str, Any]="silu" , __UpperCAmelCase : Tuple=2_048 , __UpperCAmelCase : Optional[Any]=0.02 , __UpperCAmelCase : Any=1e-6 , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : Optional[int]=0 , __UpperCAmelCase : Optional[int]=1 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : Tuple=None , **__UpperCAmelCase : Tuple , ) ->str: """simple docstring""" a = vocab_size a = max_position_embeddings a = hidden_size a = intermediate_size a = num_hidden_layers a = num_attention_heads # for backward compatibility if num_key_value_heads is None: a = num_attention_heads a = num_key_value_heads a = hidden_act a = initializer_range a = rms_norm_eps a = pretraining_tp a = use_cache a = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , tie_word_embeddings=__UpperCAmelCase , **__UpperCAmelCase , ) def __lowerCAmelCase ( self : Tuple ) ->Tuple: """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __UpperCAmelCase ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""" ) a = self.rope_scaling.get('''type''' , __UpperCAmelCase ) a = self.rope_scaling.get('''factor''' , __UpperCAmelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
0
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A : List[Any] = { 'configuration_lilt': ['LILT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LiltConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] = [ 'LILT_PRETRAINED_MODEL_ARCHIVE_LIST', 'LiltForQuestionAnswering', 'LiltForSequenceClassification', 'LiltForTokenClassification', 'LiltModel', 'LiltPreTrainedModel', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys _A : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
368
from math import factorial class __SCREAMING_SNAKE_CASE : def __init__( self : List[Any] , A : Dict , A : Any ) ->Optional[Any]: lowerCamelCase__ : Tuple = real if isinstance(A , A ): lowerCamelCase__ : Optional[int] = [1] * rank else: lowerCamelCase__ : List[Any] = rank def __repr__( self : Tuple ) ->str: return ( F"{self.real}+" F"{'+'.join(str(A )+'E'+str(n+1 )for n,dual in enumerate(self.duals ) )}" ) def __lowerCamelCase ( self : List[Any] ) ->List[Any]: lowerCamelCase__ : Tuple = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real , A ) def __add__( self : Union[str, Any] , A : int ) ->str: if not isinstance(A , A ): return Dual(self.real + other , self.duals ) lowerCamelCase__ : int = self.duals.copy() lowerCamelCase__ : int = other.duals.copy() if len(A ) > len(A ): o_dual.extend([1] * (len(A ) - len(A )) ) elif len(A ) < len(A ): s_dual.extend([1] * (len(A ) - len(A )) ) lowerCamelCase__ : Optional[Any] = [] for i in range(len(A ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real , A ) _UpperCAmelCase : List[Any] = __add__ def __sub__( self : Any , A : Dict ) ->int: return self + other * -1 def __mul__( self : Optional[Any] , A : List[Any] ) ->Union[str, Any]: if not isinstance(A , A ): lowerCamelCase__ : Optional[int] = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other , A ) lowerCamelCase__ : Tuple = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real , A ) _UpperCAmelCase : Optional[Any] = __mul__ def __truediv__( self : int , A : List[Any] ) ->Dict: if not isinstance(A , A ): lowerCamelCase__ : Tuple = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other , A ) raise ValueError def __floordiv__( self : Dict , A : Union[str, Any] ) ->Union[str, Any]: if not isinstance(A , A ): lowerCamelCase__ : Tuple = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other , A ) raise ValueError def __pow__( self : Any , A : List[Any] ) ->Tuple: if n < 0 or isinstance(A , A ): raise ValueError('''power must be a positive integer''' ) if n == 0: return 1 if n == 1: return self lowerCamelCase__ : Union[str, Any] = self for _ in range(n - 1 ): x *= self return x def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: """simple docstring""" if not callable(UpperCAmelCase ): raise ValueError('''differentiate() requires a function as input for func''' ) if not isinstance(UpperCAmelCase , (float, int) ): raise ValueError('''differentiate() requires a float as input for position''' ) if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError('''differentiate() requires an int as input for order''' ) lowerCamelCase__ : List[str] = Dual(UpperCAmelCase , 1 ) lowerCamelCase__ : Any = func(UpperCAmelCase ) if order == 0: return result.real return result.duals[order - 1] * factorial(UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() def _a ( UpperCAmelCase ) -> int: """simple docstring""" return y**2 * y**4 print(differentiate(f, 9, 2))
265
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) __UpperCamelCase : Union[str, Any] = "▁" __UpperCamelCase : Tuple = {"vocab_file": "sentencepiece.bpe.model"} __UpperCamelCase : int = { "vocab_file": { "facebook/nllb-200-distilled-600M": ( "https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model" ), } } __UpperCamelCase : Any = { "facebook/nllb-200-distilled-600M": 1024, } # fmt: off __UpperCamelCase : Optional[Any] = ["ace_Arab", "ace_Latn", "acm_Arab", "acq_Arab", "aeb_Arab", "afr_Latn", "ajp_Arab", "aka_Latn", "amh_Ethi", "apc_Arab", "arb_Arab", "ars_Arab", "ary_Arab", "arz_Arab", "asm_Beng", "ast_Latn", "awa_Deva", "ayr_Latn", "azb_Arab", "azj_Latn", "bak_Cyrl", "bam_Latn", "ban_Latn", "bel_Cyrl", "bem_Latn", "ben_Beng", "bho_Deva", "bjn_Arab", "bjn_Latn", "bod_Tibt", "bos_Latn", "bug_Latn", "bul_Cyrl", "cat_Latn", "ceb_Latn", "ces_Latn", "cjk_Latn", "ckb_Arab", "crh_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "dik_Latn", "dyu_Latn", "dzo_Tibt", "ell_Grek", "eng_Latn", "epo_Latn", "est_Latn", "eus_Latn", "ewe_Latn", "fao_Latn", "pes_Arab", "fij_Latn", "fin_Latn", "fon_Latn", "fra_Latn", "fur_Latn", "fuv_Latn", "gla_Latn", "gle_Latn", "glg_Latn", "grn_Latn", "guj_Gujr", "hat_Latn", "hau_Latn", "heb_Hebr", "hin_Deva", "hne_Deva", "hrv_Latn", "hun_Latn", "hye_Armn", "ibo_Latn", "ilo_Latn", "ind_Latn", "isl_Latn", "ita_Latn", "jav_Latn", "jpn_Jpan", "kab_Latn", "kac_Latn", "kam_Latn", "kan_Knda", "kas_Arab", "kas_Deva", "kat_Geor", "knc_Arab", "knc_Latn", "kaz_Cyrl", "kbp_Latn", "kea_Latn", "khm_Khmr", "kik_Latn", "kin_Latn", "kir_Cyrl", "kmb_Latn", "kon_Latn", "kor_Hang", "kmr_Latn", "lao_Laoo", "lvs_Latn", "lij_Latn", "lim_Latn", "lin_Latn", "lit_Latn", "lmo_Latn", "ltg_Latn", "ltz_Latn", "lua_Latn", "lug_Latn", "luo_Latn", "lus_Latn", "mag_Deva", "mai_Deva", "mal_Mlym", "mar_Deva", "min_Latn", "mkd_Cyrl", "plt_Latn", "mlt_Latn", "mni_Beng", "khk_Cyrl", "mos_Latn", "mri_Latn", "zsm_Latn", "mya_Mymr", "nld_Latn", "nno_Latn", "nob_Latn", "npi_Deva", "nso_Latn", "nus_Latn", "nya_Latn", "oci_Latn", "gaz_Latn", "ory_Orya", "pag_Latn", "pan_Guru", "pap_Latn", "pol_Latn", "por_Latn", "prs_Arab", "pbt_Arab", "quy_Latn", "ron_Latn", "run_Latn", "rus_Cyrl", "sag_Latn", "san_Deva", "sat_Beng", "scn_Latn", "shn_Mymr", "sin_Sinh", "slk_Latn", "slv_Latn", "smo_Latn", "sna_Latn", "snd_Arab", "som_Latn", "sot_Latn", "spa_Latn", "als_Latn", "srd_Latn", "srp_Cyrl", "ssw_Latn", "sun_Latn", "swe_Latn", "swh_Latn", "szl_Latn", "tam_Taml", "tat_Cyrl", "tel_Telu", "tgk_Cyrl", "tgl_Latn", "tha_Thai", "tir_Ethi", "taq_Latn", "taq_Tfng", "tpi_Latn", "tsn_Latn", "tso_Latn", "tuk_Latn", "tum_Latn", "tur_Latn", "twi_Latn", "tzm_Tfng", "uig_Arab", "ukr_Cyrl", "umb_Latn", "urd_Arab", "uzn_Latn", "vec_Latn", "vie_Latn", "war_Latn", "wol_Latn", "xho_Latn", "ydd_Hebr", "yor_Latn", "yue_Hant", "zho_Hans", "zho_Hant", "zul_Latn"] class __magic_name__ ( __lowerCAmelCase): A: str = VOCAB_FILES_NAMES A: Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A: List[Any] = PRETRAINED_VOCAB_FILES_MAP A: int = ["input_ids", "attention_mask"] A: List[int] = [] A: List[int] = [] def __init__( self : Union[str, Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any="<s>" , lowerCamelCase__ : Any="</s>" , lowerCamelCase__ : Dict="</s>" , lowerCamelCase__ : Tuple="<s>" , lowerCamelCase__ : Optional[int]="<unk>" , lowerCamelCase__ : str="<pad>" , lowerCamelCase__ : Optional[int]="<mask>" , lowerCamelCase__ : int=None , lowerCamelCase__ : List[Any]=None , lowerCamelCase__ : List[Any]=None , lowerCamelCase__ : Optional[Dict[str, Any]] = None , lowerCamelCase__ : List[Any]=None , lowerCamelCase__ : List[Any]=False , **lowerCamelCase__ : str , ) -> List[str]: '''simple docstring''' UpperCamelCase__ : Dict = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token UpperCamelCase__ : str = {} if sp_model_kwargs is None else sp_model_kwargs UpperCamelCase__ : str = legacy_behaviour super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , src_lang=lowerCamelCase__ , tgt_lang=lowerCamelCase__ , additional_special_tokens=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , legacy_behaviour=lowerCamelCase__ , **lowerCamelCase__ , ) UpperCamelCase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) UpperCamelCase__ : Optional[int] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token UpperCamelCase__ : str = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCamelCase__ : str = 1 UpperCamelCase__ : int = len(self.sp_model ) UpperCamelCase__ : str = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCamelCase__ ) } UpperCamelCase__ : Any = {v: k for k, v in self.lang_code_to_id.items()} UpperCamelCase__ : int = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCamelCase__ : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCamelCase__ : Optional[int] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCamelCase__ : Tuple = src_lang if src_lang is not None else '''eng_Latn''' UpperCamelCase__ : str = self.lang_code_to_id[self._src_lang] UpperCamelCase__ : Union[str, Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : int ) -> str: '''simple docstring''' UpperCamelCase__ : str = self.__dict__.copy() UpperCamelCase__ : Any = None UpperCamelCase__ : List[str] = self.sp_model.serialized_model_proto() return state def __setstate__( self : List[Any] , lowerCamelCase__ : int ) -> List[str]: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCamelCase__ : List[Any] = {} UpperCamelCase__ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def UpperCAmelCase__ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def UpperCAmelCase__ ( self : Union[str, Any] ) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : str ) -> None: '''simple docstring''' UpperCamelCase__ : str = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None , lowerCamelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = [1] * len(self.prefix_tokens ) UpperCamelCase__ : List[str] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCamelCase__ )) + suffix_ones return prefix_ones + ([0] * len(lowerCamelCase__ )) + ([0] * len(lowerCamelCase__ )) + suffix_ones def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' UpperCamelCase__ : Optional[int] = [self.sep_token_id] UpperCamelCase__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : str , lowerCamelCase__ : Optional[str] , lowerCamelCase__ : Optional[str] , **lowerCamelCase__ : List[Any] ) -> int: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) UpperCamelCase__ : Optional[Any] = src_lang UpperCamelCase__ : str = self(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) UpperCamelCase__ : str = self.convert_tokens_to_ids(lowerCamelCase__ ) UpperCamelCase__ : Dict = tgt_lang_id return inputs def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' UpperCamelCase__ : Tuple = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : str ) -> List[str]: '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple ) -> Any: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCamelCase__ : Optional[Any] = self.sp_model.PieceToId(lowerCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) -> Any: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Tuple = ''''''.join(lowerCamelCase__ ).replace(lowerCamelCase__ , ''' ''' ).strip() return out_string def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : str , lowerCamelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return UpperCamelCase__ : str = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , '''wb''' ) as fi: UpperCamelCase__ : Tuple = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : str = "eng_Latn" , lowerCamelCase__ : Optional[List[str]] = None , lowerCamelCase__ : str = "fra_Latn" , **lowerCamelCase__ : int , ) -> BatchEncoding: '''simple docstring''' UpperCamelCase__ : Optional[Any] = src_lang UpperCamelCase__ : Dict = tgt_lang return super().prepare_seqaseq_batch(lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def UpperCAmelCase__ ( self : str ) -> Union[str, Any]: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Union[str, Any] ) -> None: '''simple docstring''' UpperCamelCase__ : List[Any] = self.lang_code_to_id[src_lang] if self.legacy_behaviour: UpperCamelCase__ : Any = [] UpperCamelCase__ : Optional[int] = [self.eos_token_id, self.cur_lang_code] else: UpperCamelCase__ : Dict = [self.cur_lang_code] UpperCamelCase__ : Optional[int] = [self.eos_token_id] def UpperCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : str ) -> None: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = self.lang_code_to_id[lang] if self.legacy_behaviour: UpperCamelCase__ : Optional[int] = [] UpperCamelCase__ : Tuple = [self.eos_token_id, self.cur_lang_code] else: UpperCamelCase__ : List[str] = [self.cur_lang_code] UpperCamelCase__ : List[Any] = [self.eos_token_id]
146
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[int] = '''ZinengTang/tvlt-base''' UpperCamelCase__ : int = tempfile.mkdtemp() def UpperCAmelCase__ ( self : int , **lowerCamelCase__ : List[str] ) -> List[Any]: '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , **lowerCamelCase__ : Tuple ) -> List[Any]: '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : str ) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' UpperCamelCase__ : int = self.get_image_processor() UpperCamelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCamelCase__ : List[str] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase__ : Optional[int] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase__ ) self.assertIsInstance(processor.image_processor , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : str = self.get_image_processor() UpperCamelCase__ : List[Any] = self.get_feature_extractor() UpperCamelCase__ : Dict = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : Any = np.ones([12000] ) UpperCamelCase__ : Union[str, Any] = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : Any = processor(audio=lowerCamelCase__ , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = self.get_image_processor() UpperCamelCase__ : Any = self.get_feature_extractor() UpperCamelCase__ : int = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : int = np.ones([3, 224, 224] ) UpperCamelCase__ : List[str] = image_processor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : str = processor(images=lowerCamelCase__ , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Any = self.get_image_processor() UpperCamelCase__ : Dict = self.get_feature_extractor() UpperCamelCase__ : Union[str, Any] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : List[str] = np.ones([12000] ) UpperCamelCase__ : Tuple = np.ones([3, 224, 224] ) UpperCamelCase__ : Optional[Any] = processor(audio=lowerCamelCase__ , images=lowerCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(lowerCamelCase__ ): processor() def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' UpperCamelCase__ : List[str] = self.get_image_processor() UpperCamelCase__ : str = self.get_feature_extractor() UpperCamelCase__ : Tuple = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
146
1
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : float ,__lowerCamelCase : float ,__lowerCamelCase : float ,__lowerCamelCase : float ,__lowerCamelCase : float ,): lowercase_ :Union[str, Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: lowercase_ :Dict = 1 - (matter_density + radiation_density + dark_energy) lowercase_ :Dict = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) lowercase_ :Optional[Any] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation lowerCAmelCase : str =0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
147
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class a_ ( _lowerCAmelCase ): __A = ["image_processor", "tokenizer"] __A = "LayoutLMv3ImageProcessor" __A = ("LayoutLMv3Tokenizer", "LayoutLMv3TokenizerFast") def __init__( self : int , lowercase : Optional[Any]=None , lowercase : List[str]=None , **lowercase : Optional[int] ): """simple docstring""" lowercase_ :int = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase , ) lowercase_ :Optional[int] = kwargs.pop("feature_extractor" ) lowercase_ :Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase , lowercase ) def __call__( self : Optional[Any] , lowercase : List[str] , lowercase : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowercase : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , lowercase : Union[List[List[int]], List[List[List[int]]]] = None , lowercase : Optional[Union[List[int], List[List[int]]]] = None , lowercase : bool = True , lowercase : Union[bool, str, PaddingStrategy] = False , lowercase : Union[bool, str, TruncationStrategy] = None , lowercase : Optional[int] = None , lowercase : int = 0 , lowercase : Optional[int] = None , lowercase : Optional[bool] = None , lowercase : Optional[bool] = None , lowercase : bool = False , lowercase : bool = False , lowercase : bool = False , lowercase : bool = False , lowercase : bool = True , lowercase : Optional[Union[str, TensorType]] = None , **lowercase : List[Any] , ): """simple docstring""" if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) # first, apply the image processor lowercase_ :Dict = self.image_processor(images=lowercase , return_tensors=lowercase ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(lowercase , lowercase ): lowercase_ :str = [text] # add batch dimension (as the image processor always adds a batch dimension) lowercase_ :Union[str, Any] = features["words"] lowercase_ :Optional[Any] = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_token_type_ids=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) # add pixel values lowercase_ :Any = features.pop("pixel_values" ) if return_overflowing_tokens is True: lowercase_ :Any = self.get_overflowing_images(lowercase , encoded_inputs["overflow_to_sample_mapping"] ) lowercase_ :Any = images return encoded_inputs def lowercase__ ( self : List[Any] , lowercase : Any , lowercase : Optional[Any] ): """simple docstring""" lowercase_ :Union[str, Any] = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(lowercase ) != len(lowercase ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" F' {len(lowercase )} and {len(lowercase )}' ) return images_with_overflow def lowercase__ ( self : Union[str, Any] , *lowercase : List[Any] , **lowercase : Optional[Any] ): """simple docstring""" return self.tokenizer.batch_decode(*lowercase , **lowercase ) def lowercase__ ( self : List[Any] , *lowercase : Any , **lowercase : str ): """simple docstring""" return self.tokenizer.decode(*lowercase , **lowercase ) @property def lowercase__ ( self : Optional[Any] ): """simple docstring""" return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def lowercase__ ( self : str ): """simple docstring""" warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , lowercase , ) return self.image_processor_class @property def lowercase__ ( self : Tuple ): """simple docstring""" warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , lowercase , ) return self.image_processor
147
1
import cmath import math def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> complex: '''simple docstring''' UpperCamelCase = math.radians(_lowerCAmelCase ) UpperCamelCase = math.radians(_lowerCAmelCase ) # Convert voltage and current to rectangular form UpperCamelCase = cmath.rect(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase = cmath.rect(_lowerCAmelCase , _lowerCAmelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
343
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
0
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') class _a ( _lowercase , unittest.TestCase): _a : Tuple = BartphoTokenizer _a : Optional[int] = False _a : Dict = True def UpperCAmelCase__( self : str )-> str: super().setUp() lowerCAmelCase__ : Optional[int] = ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] lowerCAmelCase__ : Dict = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE ) ) ) ) lowerCAmelCase__ : List[str] = {'''unk_token''': '''<unk>'''} lowerCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''monolingual_vocab_file'''] ) with open(self.monolingual_vocab_file , '''w''' , encoding='''utf-8''' ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) lowerCAmelCase__ : List[str] = BartphoTokenizer(_SCREAMING_SNAKE_CASE , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase__( self : List[Any] , **_SCREAMING_SNAKE_CASE : Optional[int] )-> Optional[int]: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : Optional[int] )-> Dict: lowerCAmelCase__ : Any = '''This is a là test''' lowerCAmelCase__ : List[str] = '''This is a<unk><unk> test''' return input_text, output_text def UpperCAmelCase__( self : int )-> Union[str, Any]: lowerCAmelCase__ : Optional[Any] = BartphoTokenizer(_SCREAMING_SNAKE_CASE , self.monolingual_vocab_file , **self.special_tokens_map ) lowerCAmelCase__ : Union[str, Any] = '''This is a là test''' lowerCAmelCase__ : Optional[int] = '''▁This ▁is ▁a ▁l à ▁t est'''.split() lowerCAmelCase__ : Union[str, Any] = tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = tokens + [tokenizer.unk_token] lowerCAmelCase__ : List[Any] = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE )
211
from __future__ import annotations def lowerCamelCase_ ( _a , _a , _a , _a ): # noqa: E741 """simple docstring""" while r - l > 1: lowerCAmelCase__ : Any = (l + r) // 2 if v[m] >= key: lowerCAmelCase__ : int = m else: lowerCAmelCase__ : Tuple = m # noqa: E741 return r def lowerCamelCase_ ( _a ): """simple docstring""" if len(_a ) == 0: return 0 lowerCAmelCase__ : Optional[int] = [0] * len(_a ) lowerCAmelCase__ : List[Any] = 1 lowerCAmelCase__ : int = v[0] for i in range(1 , len(_a ) ): if v[i] < tail[0]: lowerCAmelCase__ : str = v[i] elif v[i] > tail[length - 1]: lowerCAmelCase__ : Any = v[i] length += 1 else: lowerCAmelCase__ : int = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
211
1
def UpperCamelCase ( snake_case__ : list[list[int]] , snake_case__ : int , snake_case__ : int , snake_case__ : set ) -> int: UpperCamelCase , UpperCamelCase : Tuple = len(snake_case__ ), len(grid[0] ) if ( min(snake_case__ , snake_case__ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) UpperCamelCase : List[Any] = 0 count += depth_first_search(snake_case__ , row + 1 , snake_case__ , snake_case__ ) count += depth_first_search(snake_case__ , row - 1 , snake_case__ , snake_case__ ) count += depth_first_search(snake_case__ , snake_case__ , col + 1 , snake_case__ ) count += depth_first_search(snake_case__ , snake_case__ , col - 1 , snake_case__ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
119
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu __UpperCAmelCase = get_tests_dir() + '''/test_data/fsmt/fsmt_val_data.json''' with io.open(filename, '''r''', encoding='''utf-8''') as f: __UpperCAmelCase = json.load(f) @require_torch class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: return FSMTTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: UpperCamelCase : List[Any] = FSMTForConditionalGeneration.from_pretrained(SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['en-ru', 26.0], ['ru-en', 22.0], ['en-de', 22.0], ['de-en', 29.0], ] ) @slow def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality UpperCamelCase : int = F"""facebook/wmt19-{pair}""" UpperCamelCase : Optional[int] = self.get_tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : str = self.get_model(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Dict = bleu_data[pair]['src'] UpperCamelCase : Tuple = bleu_data[pair]['tgt'] UpperCamelCase : List[Any] = tokenizer(SCREAMING_SNAKE_CASE_, return_tensors='pt', truncation=SCREAMING_SNAKE_CASE_, padding='longest' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Any = model.generate( input_ids=batch.input_ids, num_beams=8, ) UpperCamelCase : Tuple = tokenizer.batch_decode( SCREAMING_SNAKE_CASE_, skip_special_tokens=SCREAMING_SNAKE_CASE_, clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[Any] = calculate_bleu(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) print(SCREAMING_SNAKE_CASE_ ) self.assertGreaterEqual(scores['bleu'], SCREAMING_SNAKE_CASE_ )
119
1
import os # Precomputes a list of the 100 first triangular numbers UpperCamelCase__ = [int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = os.path.dirname(os.path.realpath(lowercase_ ) ) UpperCAmelCase__ = os.path.join(lowercase_, "words.txt" ) UpperCAmelCase__ = "" with open(lowercase_ ) as f: UpperCAmelCase__ = f.readline() UpperCAmelCase__ = [word.strip("\"" ) for word in words.strip("\r\n" ).split("," )] UpperCAmelCase__ = [ word for word in [sum(ord(lowercase_ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowercase_ ) if __name__ == "__main__": print(solution())
362
from __future__ import annotations def lowerCAmelCase_ ( __A ) -> list[int]: '''simple docstring''' if len(__A ) == 0: return array UpperCAmelCase__ , UpperCAmelCase__ = min(__A ), max(__A ) # Compute the variables UpperCAmelCase__ = _max - _min + 1 UpperCAmelCase__ , UpperCAmelCase__ = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: UpperCAmelCase__ = i - _min UpperCAmelCase__ = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. UpperCAmelCase__ = 0 for i in range(__A ): while holes_repeat[i] > 0: UpperCAmelCase__ = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = input('Enter numbers separated by comma:\n') UpperCamelCase__ = [int(x) for x in user_input.split(',')] print(pigeon_sort(unsorted))
143
0
import collections import os import re from pathlib import Path lowercase__ : Optional[Any] = "src/transformers" # Matches is_xxx_available() lowercase__ : Optional[Any] = re.compile(R"is\_([a-z_]*)_available()") # Catches a one-line _import_struct = {xxx} lowercase__ : Tuple = re.compile(R"^_import_structure\s+=\s+\{([^\}]+)\}") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowercase__ : Tuple = re.compile(R"\s+\"\S*\":\s+\[([^\]]*)\]") # Catches a line if not is_foo_available lowercase__ : List[str] = re.compile(R"^\s*if\s+not\s+is\_[a-z_]*\_available\(\)") # Catches a line _import_struct["bla"].append("foo") lowercase__ : Any = re.compile(R"^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowercase__ : str = re.compile(R"^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]") # Catches a line with an object between quotes and a comma: "MyModel", lowercase__ : int = re.compile(R"^\s+\"([^\"]+)\",") # Catches a line with objects between brackets only: ["foo", "bar"], lowercase__ : Dict = re.compile(R"^\s+\[([^\]]+)\]") # Catches a line with from foo import bar, bla, boo lowercase__ : Tuple = re.compile(R"\s+from\s+\S*\s+import\s+([^\(\s].*)\n") # Catches a line with try: lowercase__ : Dict = re.compile(R"^\s*try:") # Catches a line with else: lowercase__ : Union[str, Any] = re.compile(R"^\s*else:") def A_ ( snake_case : Optional[Any] ) -> int: '''simple docstring''' if _re_test_backend.search(snake_case ) is None: return None __UpperCamelCase = [b[0] for b in _re_backend.findall(snake_case )] backends.sort() return "_and_".join(snake_case ) def A_ ( snake_case : Tuple ) -> List[str]: '''simple docstring''' with open(snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: __UpperCamelCase = f.readlines() __UpperCamelCase = 0 while line_index < len(snake_case ) and not lines[line_index].startswith('''_import_structure = {''' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(snake_case ): return None # First grab the objects without a specific backend in _import_structure __UpperCamelCase = [] while not lines[line_index].startswith('''if TYPE_CHECKING''' ) and find_backend(lines[line_index] ) is None: __UpperCamelCase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(snake_case ): __UpperCamelCase = _re_one_line_import_struct.search(snake_case ).groups()[0] __UpperCamelCase = re.findall(r'''\[([^\]]+)\]''' , snake_case ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(''', ''' )] ) line_index += 1 continue __UpperCamelCase = _re_import_struct_key_value.search(snake_case ) if single_line_import_search is not None: __UpperCamelCase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(''', ''' ) if len(snake_case ) > 0] objects.extend(snake_case ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) line_index += 1 __UpperCamelCase = {'''none''': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('''if TYPE_CHECKING''' ): # If the line is an if not is_backend_available, we grab all objects associated. __UpperCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __UpperCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __UpperCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 4 ): __UpperCamelCase = lines[line_index] if _re_import_struct_add_one.search(snake_case ) is not None: objects.append(_re_import_struct_add_one.search(snake_case ).groups()[0] ) elif _re_import_struct_add_many.search(snake_case ) is not None: __UpperCamelCase = _re_import_struct_add_many.search(snake_case ).groups()[0].split(''', ''' ) __UpperCamelCase = [obj[1:-1] for obj in imports if len(snake_case ) > 0] objects.extend(snake_case ) elif _re_between_brackets.search(snake_case ) is not None: __UpperCamelCase = _re_between_brackets.search(snake_case ).groups()[0].split(''', ''' ) __UpperCamelCase = [obj[1:-1] for obj in imports if len(snake_case ) > 0] objects.extend(snake_case ) elif _re_quote_object.search(snake_case ) is not None: objects.append(_re_quote_object.search(snake_case ).groups()[0] ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) elif line.startswith(''' ''' * 12 + '''"''' ): objects.append(line[13:-3] ) line_index += 1 __UpperCamelCase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend __UpperCamelCase = [] while ( line_index < len(snake_case ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('''else''' ) ): __UpperCamelCase = lines[line_index] __UpperCamelCase = _re_import.search(snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 __UpperCamelCase = {'''none''': objects} # Let's continue with backend-specific objects while line_index < len(snake_case ): # If the line is an if is_backend_available, we grab all objects associated. __UpperCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __UpperCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __UpperCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 8 ): __UpperCamelCase = lines[line_index] __UpperCamelCase = _re_import.search(snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 12 ): objects.append(line[12:-2] ) line_index += 1 __UpperCamelCase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def A_ ( snake_case : List[str] , snake_case : str ) -> Dict: '''simple docstring''' def find_duplicates(snake_case : Dict ): return [k for k, v in collections.Counter(snake_case ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] __UpperCamelCase = [] for key in import_dict_objects.keys(): __UpperCamelCase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f"Duplicate _import_structure definitions for: {duplicate_imports}" ) __UpperCamelCase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f"Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): __UpperCamelCase = '''base imports''' if key == '''none''' else f"{key} backend" errors.append(f"Differences for {name}:" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f" {a} in TYPE_HINT but not in _import_structure." ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f" {a} in _import_structure but not in TYPE_HINT." ) return errors def A_ ( ) -> int: '''simple docstring''' __UpperCamelCase = [] for root, _, files in os.walk(snake_case ): if "__init__.py" in files: __UpperCamelCase = os.path.join(snake_case , '''__init__.py''' ) __UpperCamelCase = parse_init(snake_case ) if objects is not None: __UpperCamelCase = analyze_results(*snake_case ) if len(snake_case ) > 0: __UpperCamelCase = f"Problem in {fname}, both halves do not define the same objects.\n{errors[0]}" failures.append('''\n'''.join(snake_case ) ) if len(snake_case ) > 0: raise ValueError('''\n\n'''.join(snake_case ) ) def A_ ( ) -> Optional[int]: '''simple docstring''' __UpperCamelCase = [] for path, directories, files in os.walk(snake_case ): for folder in directories: # Ignore private modules if folder.startswith('''_''' ): directories.remove(snake_case ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(snake_case ) / folder).glob('''*.py''' ) ) ) == 0: continue __UpperCamelCase = str((Path(snake_case ) / folder).relative_to(snake_case ) ) __UpperCamelCase = short_path.replace(os.path.sep , '''.''' ) submodules.append(snake_case ) for fname in files: if fname == "__init__.py": continue __UpperCamelCase = str((Path(snake_case ) / fname).relative_to(snake_case ) ) __UpperCamelCase = short_path.replace('''.py''' , '''''' ).replace(os.path.sep , '''.''' ) if len(submodule.split('''.''' ) ) == 1: submodules.append(snake_case ) return submodules lowercase__ : List[str] = [ "convert_pytorch_checkpoint_to_tf2", "modeling_flax_pytorch_utils", "models.esm.openfold_utils", ] def A_ ( ) -> Any: '''simple docstring''' from transformers.utils import direct_transformers_import __UpperCamelCase = direct_transformers_import(snake_case ) __UpperCamelCase = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(snake_case , '''__init__.py''' ) , '''r''' ) as f: __UpperCamelCase = f.read() import_structure_keys.update(set(re.findall(r'''import_structure\[\"([^\"]*)\"\]''' , snake_case ) ) ) __UpperCamelCase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(snake_case ) > 0: __UpperCamelCase = '''\n'''.join(f"- {module}" for module in module_not_registered ) raise ValueError( '''The following submodules are not properly registed in the main init of Transformers:\n''' f"{list_of_modules}\n" '''Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.''' ) if __name__ == "__main__": check_all_inits() check_submodules()
328
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": lowercase__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument( "--txt2img_unclip", default="kakaobrain/karlo-v1-alpha", type=str, required=False, help="The pretrained txt2img unclip.", ) lowercase__ : Any = parser.parse_args() lowercase__ : Union[str, Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) lowercase__ : List[str] = CLIPImageProcessor() lowercase__ : Optional[Any] = CLIPVisionModelWithProjection.from_pretrained("openai/clip-vit-large-patch14") lowercase__ : Optional[Any] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
328
1
"""simple docstring""" import itertools import string from collections.abc import Generator, Iterable def snake_case_ ( A_ : Iterable[str], A_ : int ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = iter(A_ ) while True: _lowerCamelCase : Any = tuple(itertools.islice(A_, A_ ) ) if not chunk: return yield chunk def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : Tuple = ''''''.join([c.upper() for c in dirty if c in string.ascii_letters] ) _lowerCamelCase : Tuple = '''''' if len(A_ ) < 2: return dirty for i in range(len(A_ ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(A_ ) & 1: clean += "X" return clean def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : List[Any] = '''ABCDEFGHIKLMNOPQRSTUVWXYZ''' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _lowerCamelCase : List[Any] = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(A_ ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(A_ ) return table def snake_case_ ( A_ : str, A_ : str ): '''simple docstring''' _lowerCamelCase : List[str] = generate_table(A_ ) _lowerCamelCase : int = prepare_input(A_ ) _lowerCamelCase : int = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(A_, 2 ): _lowerCamelCase , _lowerCamelCase : int = divmod(table.index(A_ ), 5 ) _lowerCamelCase , _lowerCamelCase : Optional[Any] = divmod(table.index(A_ ), 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def snake_case_ ( A_ : str, A_ : str ): '''simple docstring''' _lowerCamelCase : Dict = generate_table(A_ ) _lowerCamelCase : List[Any] = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(A_, 2 ): _lowerCamelCase , _lowerCamelCase : Tuple = divmod(table.index(A_ ), 5 ) _lowerCamelCase , _lowerCamelCase : Any = divmod(table.index(A_ ), 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
175
"""simple docstring""" import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class __snake_case ( _lowercase , unittest.TestCase): snake_case__ : Optional[int] = BlenderbotSmallTokenizer snake_case__ : List[str] = False def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" super().setUp() _lowerCamelCase : str = ['''__start__''', '''adapt''', '''act''', '''ap@@''', '''te''', '''__end__''', '''__unk__'''] _lowerCamelCase : Any = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) _lowerCamelCase : Any = ['''#version: 0.2''', '''a p''', '''t e</w>''', '''ap t</w>''', '''a d''', '''ad apt</w>''', '''a c''', '''ac t</w>''', ''''''] _lowerCamelCase : List[str] = {'''unk_token''': '''__unk__''', '''bos_token''': '''__start__''', '''eos_token''': '''__end__'''} _lowerCamelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowerCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Tuple , **__lowerCAmelCase : List[str] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : List[str] ): """simple docstring""" _lowerCamelCase : int = '''adapt act apte''' _lowerCamelCase : Tuple = '''adapt act apte''' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Tuple = BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowerCamelCase : int = '''adapt act apte''' _lowerCamelCase : Optional[Any] = ['''adapt''', '''act''', '''ap@@''', '''te'''] _lowerCamelCase : Optional[Any] = tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = [tokenizer.bos_token] + tokens + [tokenizer.eos_token] _lowerCamelCase : Dict = [0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : str = BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) assert tok('''sam''' ).input_ids == [1_3_8_4] _lowerCamelCase : List[str] = '''I am a small frog.''' _lowerCamelCase : str = tok([src_text] , padding=__lowerCAmelCase , truncation=__lowerCAmelCase )['''input_ids'''] _lowerCamelCase : Any = tok.batch_decode(__lowerCAmelCase , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Union[str, Any] = BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) _lowerCamelCase : Optional[Any] = '''I am a small frog .''' _lowerCamelCase : str = '''.''' _lowerCamelCase : str = tok(__lowerCAmelCase )['''input_ids'''] _lowerCamelCase : Dict = tok(__lowerCAmelCase )['''input_ids'''] assert encoded[-1] == encoded_dot[0]
175
1
import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging lowerCAmelCase_ = logging.get_logger(__name__) def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=False ) -> Optional[int]: '''simple docstring''' try: import torch # noqa: F401 except ImportError: logger.error( '''Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise if not is_sharded: lowercase : List[str] = os.path.abspath(__magic_name__ ) logger.info(F"""Loading PyTorch weights from {pt_path}""" ) lowercase : str = torch.load(__magic_name__ , map_location='''cpu''' ) logger.info(F"""PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.""" ) lowercase : List[Any] = convert_pytorch_state_dict_to_flax(__magic_name__ , __magic_name__ ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files lowercase : Dict = convert_pytorch_sharded_state_dict_to_flax(__magic_name__ , __magic_name__ ) return flax_state_dict def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) -> (Tuple[str], np.ndarray): '''simple docstring''' def is_key_or_prefix_key_in_dict(__magic_name__ ) -> bool: return len(set(__magic_name__ ) & {key, (model_prefix,) + key} ) > 0 # layer norm lowercase : str = pt_tuple_key[:-1] + ('''scale''',) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(__magic_name__ ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean lowercase : Union[str, Any] = pt_tuple_key[:-1] + ('''mean''',) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(__magic_name__ ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var lowercase : List[Any] = pt_tuple_key[:-1] + ('''var''',) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(__magic_name__ ): return renamed_pt_tuple_key, pt_tensor # embedding lowercase : Optional[Any] = pt_tuple_key[:-1] + ('''embedding''',) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(__magic_name__ ): return renamed_pt_tuple_key, pt_tensor # conv layer lowercase : Union[str, Any] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(__magic_name__ ): lowercase : Optional[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer lowercase : Union[str, Any] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(__magic_name__ ): lowercase : Optional[int] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight lowercase : Union[str, Any] = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias lowercase : Union[str, Any] = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 lowercase : str = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): lowercase : List[Any] = pt_tuple_key[-2] + '''_g''' elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): lowercase : str = pt_tuple_key[-2] + '''_v''' if name is not None: lowercase : int = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def snake_case( __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' lowercase : Optional[Any] = {k: v.numpy() for k, v in pt_state_dict.items()} lowercase : Any = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: lowercase : str = flax_model.params['''params'''] else: lowercase : Optional[int] = flax_model.params lowercase : Tuple = flatten_dict(__magic_name__ ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: lowercase : Tuple = flatten_dict(flax_model.params['''batch_stats'''] ) random_flax_state_dict.update(__magic_name__ ) lowercase : List[str] = {} lowercase : Union[str, Any] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('''.''' )[0] for k in pt_state_dict.keys()} ) lowercase : Dict = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('''.''' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowercase : Optional[Any] = tuple(pt_key.split('''.''' ) ) # remove base model prefix if necessary lowercase : str = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: lowercase : int = pt_tuple_key[1:] # Correctly rename weight parameters lowercase , lowercase : Union[str, Any] = rename_key_and_reshape_tensor( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # add model prefix if necessary lowercase : Tuple = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: lowercase : str = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: lowercase : Union[str, Any] = jnp.asarray(__magic_name__ ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(__magic_name__ , __magic_name__ ) continue # also add unexpected weight so that warning is thrown lowercase : Optional[int] = jnp.asarray(__magic_name__ ) else: # also add unexpected weight so that warning is thrown lowercase : int = jnp.asarray(__magic_name__ ) return unflatten_dict(__magic_name__ ) def snake_case( __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' import torch # Load the index lowercase : Dict = {} for shard_file in shard_filenames: # load using msgpack utils lowercase : List[str] = torch.load(__magic_name__ ) lowercase : Optional[int] = {k: v.numpy() for k, v in pt_state_dict.items()} lowercase : Any = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: lowercase : str = flax_model.params['''params'''] lowercase : str = flatten_dict(__magic_name__ ) random_flax_state_dict.update(flatten_dict(flax_model.params['''batch_stats'''] ) ) else: lowercase : int = flax_model.params lowercase : Optional[int] = flatten_dict(__magic_name__ ) lowercase : Optional[int] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('''.''' )[0] for k in pt_state_dict.keys()} ) lowercase : int = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('''.''' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowercase : Tuple = tuple(pt_key.split('''.''' ) ) # remove base model prefix if necessary lowercase : Optional[int] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: lowercase : List[Any] = pt_tuple_key[1:] # Correctly rename weight parameters lowercase , lowercase : Optional[Any] = rename_key_and_reshape_tensor( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # add model prefix if necessary lowercase : Optional[Any] = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: lowercase : List[str] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: lowercase : List[str] = jnp.asarray(__magic_name__ ) continue if "var" in flax_key[-1]: lowercase : Optional[int] = jnp.asarray(__magic_name__ ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(__magic_name__ , __magic_name__ ) continue # also add unexpected weight so that warning is thrown lowercase : Optional[Any] = jnp.asarray(__magic_name__ ) else: # also add unexpected weight so that warning is thrown lowercase : Tuple = jnp.asarray(__magic_name__ ) return unflatten_dict(__magic_name__ ) def snake_case( __magic_name__ , __magic_name__ ) -> Any: '''simple docstring''' lowercase : Optional[Any] = os.path.abspath(__magic_name__ ) logger.info(F"""Loading Flax weights from {flax_checkpoint_path}""" ) # import correct flax class lowercase : str = getattr(__magic_name__ , '''Flax''' + model.__class__.__name__ ) # load flax weight dict with open(__magic_name__ , '''rb''' ) as state_f: try: lowercase : Optional[Any] = from_bytes(__magic_name__ , state_f.read() ) except UnpicklingError: raise EnvironmentError(F"""Unable to convert {flax_checkpoint_path} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(__magic_name__ , __magic_name__ ) def snake_case( __magic_name__ , __magic_name__ ) -> int: '''simple docstring''' try: import torch # noqa: F401 except ImportError: logger.error( '''Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights lowercase : Optional[Any] = flatten_dict(jax.tree_util.tree_map(lambda __magic_name__ : x.dtype == jnp.bfloataa , __magic_name__ ) ).values() if any(__magic_name__ ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) lowercase : str = jax.tree_util.tree_map( lambda __magic_name__ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , __magic_name__ ) lowercase : int = flatten_dict(__magic_name__ ) lowercase : Tuple = pt_model.state_dict() lowercase : Dict = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split('''.''' )[0] for k in pt_model_dict.keys()} ) lowercase : Optional[int] = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split('''.''' )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys lowercase : Tuple = [] lowercase : int = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowercase : int = flax_key_tuple[0] == pt_model.base_model_prefix lowercase : List[str] = '''.'''.join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: lowercase : List[Any] = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: lowercase : str = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(__magic_name__ ) not in pt_model_dict: # conv layer lowercase : Optional[int] = flax_key_tuple[:-1] + ('''weight''',) lowercase : int = jnp.transpose(__magic_name__ , (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(__magic_name__ ) not in pt_model_dict: # linear layer lowercase : Dict = flax_key_tuple[:-1] + ('''weight''',) lowercase : Union[str, Any] = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: lowercase : List[str] = flax_key_tuple[:-1] + ('''weight''',) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: lowercase : Any = flax_key_tuple[:-1] + ('''running_mean''',) elif "var" in flax_key_tuple[-1]: lowercase : Any = flax_key_tuple[:-1] + ('''running_var''',) if "batch_stats" in flax_state: lowercase : Any = '''.'''.join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: lowercase : Union[str, Any] = '''.'''.join(__magic_name__ ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. lowercase : Tuple = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: lowercase : Dict = key.split('''.''' ) lowercase : List[str] = None if key_components[-3::2] == ["parametrizations", "original0"]: lowercase : Any = key_components[-2] + '''_g''' elif key_components[-3::2] == ["parametrizations", "original1"]: lowercase : Optional[int] = key_components[-2] + '''_v''' if name is not None: lowercase : List[Any] = key_components[:-3] + [name] lowercase : str = '''.'''.join(__magic_name__ ) lowercase : Any = key if flax_key in special_pt_names: lowercase : List[Any] = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict lowercase : Tuple = np.asarray(__magic_name__ ) if not isinstance(__magic_name__ , np.ndarray ) else flax_tensor lowercase : Optional[Any] = torch.from_numpy(__magic_name__ ) # remove from missing keys missing_keys.remove(__magic_name__ ) else: # weight is not expected by PyTorch model unexpected_keys.append(__magic_name__ ) pt_model.load_state_dict(__magic_name__ ) # re-transform missing_keys to list lowercase : Any = list(__magic_name__ ) if len(__magic_name__ ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) else: logger.warning(F"""All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n""" ) if len(__magic_name__ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" ''' use it for predictions and inference.''' ) else: logger.warning( F"""All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n""" '''If your task is similar to the task the model of the checkpoint was trained on, ''' F"""you can already use {pt_model.__class__.__name__} for predictions without further training.""" ) return pt_model
308
def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' lowercase : Union[str, Any] = [False] * len(__magic_name__ ) lowercase : Optional[int] = [] queue.append(__magic_name__ ) lowercase : int = True while queue: lowercase : Union[str, Any] = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__magic_name__ ) lowercase : Dict = True lowercase : List[str] = u return visited[t] def snake_case( __magic_name__ , __magic_name__ , __magic_name__ ) -> Tuple: '''simple docstring''' lowercase : List[str] = [-1] * (len(__magic_name__ )) lowercase : Tuple = 0 while bfs(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase : Any = float('''Inf''' ) lowercase : str = sink while s != source: # Find the minimum value in select path lowercase : Any = min(__magic_name__ , graph[parent[s]][s] ) lowercase : Dict = parent[s] max_flow += path_flow lowercase : Union[str, Any] = sink while v != source: lowercase : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow lowercase : Optional[int] = parent[v] return max_flow lowerCAmelCase_ = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] lowerCAmelCase_ , lowerCAmelCase_ = 0, 5 print(ford_fulkerson(graph, source, sink))
308
1
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a : Optional[int] = logging.get_logger(__name__) _a : Dict = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Optional[Any] = "detr" _UpperCamelCase : Union[str, Any] = ["past_key_values"] _UpperCamelCase : Any = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , a__=True , a__=None , a__=3 , a__=100 , a__=6 , a__=2048 , a__=8 , a__=6 , a__=2048 , a__=8 , a__=0.0 , a__=0.0 , a__=True , a__="relu" , a__=256 , a__=0.1 , a__=0.0 , a__=0.0 , a__=0.0_2 , a__=1.0 , a__=False , a__="sine" , a__="resnet50" , a__=True , a__=False , a__=1 , a__=5 , a__=2 , a__=1 , a__=1 , a__=5 , a__=2 , a__=0.1 , **a__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) _lowerCAmelCase : Dict = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(a__ , a__ ): _lowerCAmelCase : str = backbone_config.get("""model_type""" ) _lowerCAmelCase : Any = CONFIG_MAPPING[backbone_model_type] _lowerCAmelCase : Union[str, Any] = config_class.from_dict(a__ ) # set timm attributes to None _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Dict = None, None, None _lowerCAmelCase : Optional[int] = use_timm_backbone _lowerCAmelCase : Union[str, Any] = backbone_config _lowerCAmelCase : Tuple = num_channels _lowerCAmelCase : List[str] = num_queries _lowerCAmelCase : str = d_model _lowerCAmelCase : int = encoder_ffn_dim _lowerCAmelCase : Union[str, Any] = encoder_layers _lowerCAmelCase : List[Any] = encoder_attention_heads _lowerCAmelCase : Optional[Any] = decoder_ffn_dim _lowerCAmelCase : Tuple = decoder_layers _lowerCAmelCase : Dict = decoder_attention_heads _lowerCAmelCase : Optional[Any] = dropout _lowerCAmelCase : str = attention_dropout _lowerCAmelCase : Optional[int] = activation_dropout _lowerCAmelCase : Dict = activation_function _lowerCAmelCase : Any = init_std _lowerCAmelCase : Dict = init_xavier_std _lowerCAmelCase : int = encoder_layerdrop _lowerCAmelCase : int = decoder_layerdrop _lowerCAmelCase : Tuple = encoder_layers _lowerCAmelCase : List[Any] = auxiliary_loss _lowerCAmelCase : Optional[int] = position_embedding_type _lowerCAmelCase : List[str] = backbone _lowerCAmelCase : str = use_pretrained_backbone _lowerCAmelCase : Any = dilation # Hungarian matcher _lowerCAmelCase : List[str] = class_cost _lowerCAmelCase : Union[str, Any] = bbox_cost _lowerCAmelCase : List[str] = giou_cost # Loss coefficients _lowerCAmelCase : List[Any] = mask_loss_coefficient _lowerCAmelCase : Optional[int] = dice_loss_coefficient _lowerCAmelCase : int = bbox_loss_coefficient _lowerCAmelCase : Dict = giou_loss_coefficient _lowerCAmelCase : str = eos_coefficient super().__init__(is_encoder_decoder=a__ , **a__ ) @property def __A ( self ): return self.encoder_attention_heads @property def __A ( self ): return self.d_model @classmethod def __A ( cls , a__ , **a__ ): return cls(backbone_config=a__ , **a__ ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: _lowerCAmelCase : Dict = self.backbone_config.to_dict() _lowerCAmelCase : Any = self.__class__.model_type return output class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Optional[int] = version.parse("1.11" ) @property def __A ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def __A ( self ): return 1e-5 @property def __A ( self ): return 12
126
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Any = logging.get_logger(__name__) _a : Any = { 'uclanlp/visualbert-vqa': 'https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json', 'uclanlp/visualbert-vqa-pre': 'https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json', 'uclanlp/visualbert-vqa-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-vcr': 'https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json', 'uclanlp/visualbert-vcr-pre': 'https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json', 'uclanlp/visualbert-vcr-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-nlvr2': 'https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-pre': 'https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : List[Any] = "visual_bert" def __init__( self , a__=30522 , a__=768 , a__=512 , a__=12 , a__=12 , a__=3072 , a__="gelu" , a__=0.1 , a__=0.1 , a__=512 , a__=2 , a__=0.0_2 , a__=1e-12 , a__=False , a__=True , a__=1 , a__=0 , a__=2 , **a__ , ): super().__init__(pad_token_id=a__ , bos_token_id=a__ , eos_token_id=a__ , **a__ ) _lowerCAmelCase : Optional[Any] = vocab_size _lowerCAmelCase : int = max_position_embeddings _lowerCAmelCase : Optional[Any] = hidden_size _lowerCAmelCase : Optional[int] = visual_embedding_dim _lowerCAmelCase : Optional[int] = num_hidden_layers _lowerCAmelCase : List[Any] = num_attention_heads _lowerCAmelCase : Any = intermediate_size _lowerCAmelCase : List[Any] = hidden_act _lowerCAmelCase : int = hidden_dropout_prob _lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob _lowerCAmelCase : Any = initializer_range _lowerCAmelCase : List[Any] = type_vocab_size _lowerCAmelCase : Union[str, Any] = layer_norm_eps _lowerCAmelCase : Optional[Any] = bypass_transformer _lowerCAmelCase : List[Any] = special_visual_initialize
126
1
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __a = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt') __a = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) __a = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowercase__: """simple docstring""" a :Optional[str] = field( default='cifar10' , metadata={'help': 'Name of a dataset from the datasets package'} ) a :Optional[str] = field( default=UpperCAmelCase , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) a :Optional[str] = field( default=UpperCAmelCase , metadata={'help': 'The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'} , ) a :Optional[str] = field(default=UpperCAmelCase , metadata={'help': 'A folder containing the training data.'} ) a :Optional[str] = field(default=UpperCAmelCase , metadata={'help': 'A folder containing the validation data.'} ) a :Optional[float] = field( default=0.15 , metadata={'help': 'Percent to split off of train for validation.'} ) a :int = field(default=32 , metadata={'help': 'The size of the square patches to use for masking.'} ) a :float = field( default=0.6 , metadata={'help': 'Percentage of patches to mask.'} , ) a :Optional[int] = field( default=UpperCAmelCase , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) a :Optional[int] = field( default=UpperCAmelCase , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def _lowercase ( self : List[str] ) -> Tuple: lowercase_ = {} if self.train_dir is not None: lowercase_ = self.train_dir if self.validation_dir is not None: lowercase_ = self.validation_dir lowercase_ = data_files if data_files else None @dataclass class lowercase__: """simple docstring""" a :str = field( default=UpperCAmelCase , metadata={ 'help': ( 'The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ' 'checkpoint identifier on the hub. ' 'Don\'t set if you want to train a model from scratch.' ) } , ) a :Optional[str] = field( default=UpperCAmelCase , metadata={'help': 'If training from scratch, pass a model type from the list: ' + ', '.join(UpperCAmelCase )} , ) a :Optional[str] = field( default=UpperCAmelCase , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) a :Optional[str] = field( default=UpperCAmelCase , metadata={ 'help': ( 'Override some existing default config settings when a model is trained from scratch. Example: ' 'n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index' ) } , ) a :Optional[str] = field( default=UpperCAmelCase , metadata={'help': 'Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'} , ) a :str = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) a :str = field(default=UpperCAmelCase , metadata={'help': 'Name or path of preprocessor config.'} ) a :bool = field( default=UpperCAmelCase , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) a :Optional[int] = field( default=UpperCAmelCase , metadata={ 'help': ( 'The size (resolution) of each image. If not specified, will use `image_size` of the configuration.' ) } , ) a :Optional[int] = field( default=UpperCAmelCase , metadata={ 'help': ( 'The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.' ) } , ) a :Optional[int] = field( default=UpperCAmelCase , metadata={'help': 'Stride to use for the encoder.'} , ) class lowercase__: """simple docstring""" def __init__( self : Any , SCREAMING_SNAKE_CASE_ : List[str]=1_9_2 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=3_2 , SCREAMING_SNAKE_CASE_ : Tuple=4 , SCREAMING_SNAKE_CASE_ : List[str]=0.6 ) -> List[str]: lowercase_ = input_size lowercase_ = mask_patch_size lowercase_ = model_patch_size lowercase_ = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError('''Input size must be divisible by mask patch size''' ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError('''Mask patch size must be divisible by model patch size''' ) lowercase_ = self.input_size // self.mask_patch_size lowercase_ = self.mask_patch_size // self.model_patch_size lowercase_ = self.rand_size**2 lowercase_ = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self : List[str] ) -> List[Any]: lowercase_ = np.random.permutation(self.token_count )[: self.mask_count] lowercase_ = np.zeros(self.token_count , dtype=SCREAMING_SNAKE_CASE_ ) lowercase_ = 1 lowercase_ = mask.reshape((self.rand_size, self.rand_size) ) lowercase_ = mask.repeat(self.scale , axis=0 ).repeat(self.scale , axis=1 ) return torch.tensor(mask.flatten() ) def a ( snake_case__: str ): '''simple docstring''' lowercase_ = torch.stack([example['''pixel_values'''] for example in examples] ) lowercase_ = torch.stack([example['''mask'''] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def a ( ): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowercase_ , lowercase_ , lowercase_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowercase_ , lowercase_ , lowercase_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_mim''' , snake_case__ , snake_case__ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowercase_ = training_args.get_process_log_level() logger.setLevel(snake_case__ ) transformers.utils.logging.set_verbosity(snake_case__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. lowercase_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset. lowercase_ = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. lowercase_ = None if '''validation''' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , snake_case__ ) and data_args.train_val_split > 0.0: lowercase_ = ds['''train'''].train_test_split(data_args.train_val_split ) lowercase_ = split['''train'''] lowercase_ = split['''test'''] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase_ = { '''cache_dir''': model_args.cache_dir, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.config_name_or_path: lowercase_ = AutoConfig.from_pretrained(model_args.config_name_or_path , **snake_case__ ) elif model_args.model_name_or_path: lowercase_ = AutoConfig.from_pretrained(model_args.model_name_or_path , **snake_case__ ) else: lowercase_ = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(snake_case__ , '''decoder_type''' ): lowercase_ = '''simmim''' # adapt config lowercase_ = model_args.image_size if model_args.image_size is not None else config.image_size lowercase_ = model_args.patch_size if model_args.patch_size is not None else config.patch_size lowercase_ = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { '''image_size''': model_args.image_size, '''patch_size''': model_args.patch_size, '''encoder_stride''': model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: lowercase_ = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **snake_case__ ) elif model_args.model_name_or_path: lowercase_ = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **snake_case__ ) else: lowercase_ = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } lowercase_ = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: lowercase_ = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=snake_case__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('''Training new model from scratch''' ) lowercase_ = AutoModelForMaskedImageModeling.from_config(snake_case__ ) if training_args.do_train: lowercase_ = ds['''train'''].column_names else: lowercase_ = ds['''validation'''].column_names if data_args.image_column_name is not None: lowercase_ = data_args.image_column_name elif "image" in column_names: lowercase_ = '''image''' elif "img" in column_names: lowercase_ = '''img''' else: lowercase_ = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py lowercase_ = Compose( [ Lambda(lambda snake_case__ : img.convert('''RGB''' ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.6_7, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator lowercase_ = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(snake_case__: Optional[int] ): lowercase_ = [transforms(snake_case__ ) for image in examples[image_column_name]] lowercase_ = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError('''--do_train requires a train dataset''' ) if data_args.max_train_samples is not None: lowercase_ = ds['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(snake_case__ ) if training_args.do_eval: if "validation" not in ds: raise ValueError('''--do_eval requires a validation dataset''' ) if data_args.max_eval_samples is not None: lowercase_ = ( ds['''validation'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(snake_case__ ) # Initialize our trainer lowercase_ = Trainer( model=snake_case__ , args=snake_case__ , train_dataset=ds['''train'''] if training_args.do_train else None , eval_dataset=ds['''validation'''] if training_args.do_eval else None , tokenizer=snake_case__ , data_collator=snake_case__ , ) # Training if training_args.do_train: lowercase_ = None if training_args.resume_from_checkpoint is not None: lowercase_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowercase_ = last_checkpoint lowercase_ = trainer.train(resume_from_checkpoint=snake_case__ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: lowercase_ = trainer.evaluate() trainer.log_metrics('''eval''' , snake_case__ ) trainer.save_metrics('''eval''' , snake_case__ ) # Write model card and (optionally) push to hub lowercase_ = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''masked-image-modeling''', '''dataset''': data_args.dataset_name, '''tags''': ['''masked-image-modeling'''], } if training_args.push_to_hub: trainer.push_to_hub(**snake_case__ ) else: trainer.create_model_card(**snake_case__ ) if __name__ == "__main__": main()
30
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _UpperCAmelCase : int = logging.get_logger(__name__) # pylint: disable=invalid-name _UpperCAmelCase : Dict = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase=8 ): lowercase :List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase :List[Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class __lowerCAmelCase ( lowerCAmelCase): def __init__( self: Tuple , _lowerCAmelCase: UNetaDConditionModel , _lowerCAmelCase: DDPMScheduler , _lowerCAmelCase: VQModel , ): super().__init__() self.register_modules( unet=_lowerCAmelCase , scheduler=_lowerCAmelCase , movq=_lowerCAmelCase , ) lowercase :List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def SCREAMING_SNAKE_CASE ( self: List[str] , _lowerCAmelCase: Dict , _lowerCAmelCase: Optional[Any] , _lowerCAmelCase: Any , _lowerCAmelCase: Tuple , _lowerCAmelCase: Tuple , _lowerCAmelCase: Optional[int] ): if latents is None: lowercase :int = randn_tensor(_lowerCAmelCase , generator=_lowerCAmelCase , device=_lowerCAmelCase , dtype=_lowerCAmelCase ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) lowercase :Optional[Any] = latents.to(_lowerCAmelCase ) lowercase :int = latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: List[str]=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase :List[Any] = torch.device(F"cuda:{gpu_id}" ) lowercase :List[str] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowerCAmelCase , _lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self: List[str] , _lowerCAmelCase: Dict=0 ): if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase :List[Any] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowerCAmelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase :List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase , lowercase :Dict = cpu_offload_with_hook(_lowerCAmelCase , _lowerCAmelCase , prev_module_hook=_lowerCAmelCase ) # We'll offload the last model manually. lowercase :Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE ( self: int ): if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowerCAmelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowerCAmelCase ) def __call__( self: str , _lowerCAmelCase: Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowerCAmelCase: Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowerCAmelCase: int = 5_12 , _lowerCAmelCase: int = 5_12 , _lowerCAmelCase: int = 1_00 , _lowerCAmelCase: float = 4.0 , _lowerCAmelCase: int = 1 , _lowerCAmelCase: Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowerCAmelCase: Optional[torch.FloatTensor] = None , _lowerCAmelCase: Optional[str] = "pil" , _lowerCAmelCase: bool = True , ): lowercase :str = self._execution_device lowercase :List[str] = guidance_scale > 1.0 if isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowercase :Any = torch.cat(_lowerCAmelCase , dim=0 ) lowercase :Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowercase :List[Any] = torch.cat(_lowerCAmelCase , dim=0 ) if do_classifier_free_guidance: lowercase :int = image_embeds.repeat_interleave(_lowerCAmelCase , dim=0 ) lowercase :Any = negative_image_embeds.repeat_interleave(_lowerCAmelCase , dim=0 ) lowercase :Optional[int] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowerCAmelCase ) self.scheduler.set_timesteps(_lowerCAmelCase , device=_lowerCAmelCase ) lowercase :Optional[Any] = self.scheduler.timesteps lowercase :Tuple = self.unet.config.in_channels lowercase , lowercase :Optional[int] = downscale_height_and_width(_lowerCAmelCase , _lowerCAmelCase , self.movq_scale_factor ) # create initial latent lowercase :List[Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowerCAmelCase ) ): # expand the latents if we are doing classifier free guidance lowercase :Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase :List[str] = {"image_embeds": image_embeds} lowercase :List[Any] = self.unet( sample=_lowerCAmelCase , timestep=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , added_cond_kwargs=_lowerCAmelCase , return_dict=_lowerCAmelCase , )[0] if do_classifier_free_guidance: lowercase , lowercase :List[Any] = noise_pred.split(latents.shape[1] , dim=1 ) lowercase , lowercase :Any = noise_pred.chunk(2 ) lowercase , lowercase :Dict = variance_pred.chunk(2 ) lowercase :str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase :Any = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase , lowercase :str = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase :Tuple = self.scheduler.step( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , generator=_lowerCAmelCase , )[0] # post-processing lowercase :Dict = self.movq.decode(_lowerCAmelCase , force_not_quantize=_lowerCAmelCase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: lowercase :Any = image * 0.5 + 0.5 lowercase :Tuple = image.clamp(0 , 1 ) lowercase :List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase :List[Any] = self.numpy_to_pil(_lowerCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowerCAmelCase )
236
0
from datetime import datetime as dt import os from github import Github __UpperCAmelCase : int = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def A__ ( ) -> str: __snake_case: Dict = Github(os.environ["""GITHUB_TOKEN"""]) __snake_case: Any = g.get_repo("""huggingface/transformers""") __snake_case: Any = repo.get_issues(state="""open""") for issue in open_issues: __snake_case: Union[str, Any] = sorted([comment for comment in issue.get_comments()] , key=lambda SCREAMING_SNAKE_CASE__: i.created_at , reverse=A__) __snake_case: Union[str, Any] = comments[0] if len(A__) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""") if __name__ == "__main__": main()
351
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase : str = logging.get_logger(__name__) __UpperCAmelCase : int = { "RWKV/rwkv-4-169m-pile": "https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json", "RWKV/rwkv-4-430m-pile": "https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json", "RWKV/rwkv-4-1b5-pile": "https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json", "RWKV/rwkv-4-3b-pile": "https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json", "RWKV/rwkv-4-7b-pile": "https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json", "RWKV/rwkv-4-14b-pile": "https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json", "RWKV/rwkv-raven-1b5": "https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json", "RWKV/rwkv-raven-3b": "https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json", "RWKV/rwkv-raven-7b": "https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json", "RWKV/rwkv-raven-14b": "https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json", } class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = """rwkv""" lowerCAmelCase__ = {"""max_position_embeddings""": """context_length"""} def __init__( self : Dict , A : List[Any]=50_277 , A : List[Any]=1_024 , A : Union[str, Any]=4_096 , A : Tuple=32 , A : List[Any]=None , A : Tuple=None , A : Tuple=1E-5 , A : int=0 , A : Optional[int]=0 , A : Dict=6 , A : Dict=False , A : int=True , **A : List[Any] , ): __snake_case: Tuple = vocab_size __snake_case: Any = context_length __snake_case: Dict = hidden_size __snake_case: Dict = num_hidden_layers __snake_case: Union[str, Any] = attention_hidden_size if attention_hidden_size is not None else hidden_size __snake_case: str = intermediate_size if intermediate_size is not None else 4 * hidden_size __snake_case: Any = layer_norm_epsilon __snake_case: int = rescale_every __snake_case: str = use_cache __snake_case: Dict = bos_token_id __snake_case: Union[str, Any] = eos_token_id super().__init__( tie_word_embeddings=A , bos_token_id=A , eos_token_id=A , **A )
293
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase: Optional[Any] = logging.get_logger(__name__) def lowerCamelCase__ ( _A ): a : Optional[Any] = 'huggingface/label-files' a : Tuple = 'imagenet-1k-id2label.json' a : int = json.load(open(hf_hub_download(A__ , A__ , repo_type='dataset' ) , 'r' ) ) a : Union[str, Any] = {int(A__ ): v for k, v in idalabel.items()} a : Dict = {v: k for k, v in idalabel.items()} a : Optional[Any] = 'std_conv' if 'bit' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" a : str = BitConfig( conv_layer=A__ , num_labels=1000 , idalabel=A__ , labelaid=A__ , ) return config def lowerCamelCase__ ( _A ): if "stem.conv" in name: a : Dict = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: a : Optional[Any] = name.replace('blocks' , 'layers' ) if "head.fc" in name: a : Tuple = name.replace('head.fc' , 'classifier.1' ) if name.startswith('norm' ): a : Optional[Any] = 'bit.' + name if "bit" not in name and "classifier" not in name: a : str = 'bit.encoder.' + name return name def lowerCamelCase__ ( ): a : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' a : Union[str, Any] = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( _A , _A , _A=False ): a : Any = get_config(A__ ) # load original model from timm a : Dict = create_model(A__ , pretrained=A__ ) timm_model.eval() # load state_dict of original model a : Dict = timm_model.state_dict() for key in state_dict.copy().keys(): a : Tuple = state_dict.pop(A__ ) a : Optional[int] = val.squeeze() if 'head' in key else val # load HuggingFace model a : Optional[Any] = BitForImageClassification(A__ ) model.eval() model.load_state_dict(A__ ) # create image processor a : Optional[Any] = create_transform(**resolve_data_config({} , model=A__ ) ) a : List[str] = transform.transforms a : Tuple = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } a : Dict = BitImageProcessor( do_resize=A__ , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=A__ , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=A__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) a : List[str] = prepare_img() a : Union[str, Any] = transform(A__ ).unsqueeze(0 ) a : List[Any] = processor(A__ , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(A__ , A__ ) # verify logits with torch.no_grad(): a : List[str] = model(A__ ) a : Dict = outputs.logits print('Logits:' , logits[0, :3] ) print('Predicted class:' , model.config.idalabel[logits.argmax(-1 ).item()] ) a : Union[str, Any] = timm_model(A__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(A__ , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(A__ ).mkdir(exist_ok=A__ ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCAmelCase: Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='resnetv2_50x1_bitm', type=str, help='Name of the BiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model to the hub.', ) lowerCAmelCase: Tuple = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
297
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCamelCase : List[Any] = { "configuration_m2m_100": ["M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP", "M2M100Config", "M2M100OnnxConfig"], "tokenization_m2m_100": ["M2M100Tokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : int = [ "M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST", "M2M100ForConditionalGeneration", "M2M100Model", "M2M100PreTrainedModel", ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys _lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
28
0
import cmath import math def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : List[Any] = math.radians(_lowerCamelCase ) _lowerCAmelCase : Optional[Any] = math.radians(_lowerCamelCase ) # Convert voltage and current to rectangular form _lowerCAmelCase : Tuple = cmath.rect(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : List[Any] = cmath.rect(_lowerCamelCase , _lowerCamelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
300
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _snake_case = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ["ConvNextFeatureExtractor"] _snake_case = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _snake_case = _LazyModule(__name__, globals()["__file__"], _import_structure)
300
1
"""simple docstring""" from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ["image_processor", "tokenizer"] snake_case__ = "Pix2StructImageProcessor" snake_case__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Union[str, Any] ): UpperCAmelCase__ = False super().__init__(lowerCamelCase__ ,lowerCamelCase__ ) def __call__( self : Optional[int] ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCamelCase__ : Union[bool, str, TruncationStrategy] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = 2_048 ,lowerCamelCase__ : int = 0 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,**lowerCamelCase__ : Union[str, Any] ,): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: UpperCAmelCase__ = self.tokenizer UpperCAmelCase__ = self.tokenizer( text=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,padding=lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=lowerCamelCase__ ,stride=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_overflowing_tokens=lowerCamelCase__ ,return_special_tokens_mask=lowerCamelCase__ ,return_offsets_mapping=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,return_length=lowerCamelCase__ ,verbose=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ,) return text_encoding if not self.image_processor.is_vqa: # add pixel_values UpperCAmelCase__ = self.image_processor( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,max_patches=lowerCamelCase__ ,**lowerCamelCase__ ) else: # add pixel_values and bbox UpperCAmelCase__ = self.image_processor( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,max_patches=lowerCamelCase__ ,header_text=lowerCamelCase__ ,**lowerCamelCase__ ) if text is not None and not self.image_processor.is_vqa: UpperCAmelCase__ = self.tokenizer( text=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,padding=lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=lowerCamelCase__ ,stride=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_overflowing_tokens=lowerCamelCase__ ,return_special_tokens_mask=lowerCamelCase__ ,return_offsets_mapping=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,return_length=lowerCamelCase__ ,verbose=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ,) if "attention_mask" in text_encoding: UpperCAmelCase__ = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: UpperCAmelCase__ = text_encoding.pop('input_ids' ) else: UpperCAmelCase__ = None if text_encoding is not None: encoding_image_processor.update(lowerCamelCase__ ) return encoding_image_processor def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : Dict ,**lowerCamelCase__ : Union[str, Any] ): return self.tokenizer.batch_decode(*lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : str ,**lowerCamelCase__ : Tuple ): return self.tokenizer.decode(*lowerCamelCase__ ,**lowerCamelCase__ ) @property def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.tokenizer.model_input_names UpperCAmelCase__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
98
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCAmelCase__ : str = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[str] = [ 'ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoCBertForCausalLM', 'RoCBertForMaskedLM', 'RoCBertForMultipleChoice', 'RoCBertForPreTraining', 'RoCBertForQuestionAnswering', 'RoCBertForSequenceClassification', 'RoCBertForTokenClassification', 'RoCBertLayer', 'RoCBertModel', 'RoCBertPreTrainedModel', 'load_tf_weights_in_roc_bert', ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys lowerCAmelCase__ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
98
1
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class a_ ( a__ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = BertJapaneseTokenizer __SCREAMING_SNAKE_CASE : int = False __SCREAMING_SNAKE_CASE : Tuple = True def __lowerCAmelCase ( self ) ->List[str]: super().setUp() SCREAMING_SNAKE_CASE : Optional[int] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] SCREAMING_SNAKE_CASE : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __lowerCAmelCase ( self , _lowerCamelCase ) ->Tuple: SCREAMING_SNAKE_CASE : Tuple = '''こんにちは、世界。 \nこんばんは、世界。''' SCREAMING_SNAKE_CASE : str = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __lowerCAmelCase ( self , _lowerCamelCase ) ->Tuple: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = self.get_input_output_texts(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase ) SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.decode(_lowerCamelCase , clean_up_tokenization_spaces=_lowerCamelCase ) return text, ids def __lowerCAmelCase ( self ) ->Tuple: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Optional[int]: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Optional[Any]: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Dict: SCREAMING_SNAKE_CASE : str = self.tokenizer_class(self.vocab_file ) SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(_lowerCamelCase , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Dict = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Union[str, Any] = '''こんにちは、世界。\nこんばんは、世界。''' SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) SCREAMING_SNAKE_CASE : Tuple = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_lowerCamelCase , '''wb''' ) as handle: pickle.dump(_lowerCamelCase , _lowerCamelCase ) with open(_lowerCamelCase , '''rb''' ) as handle: SCREAMING_SNAKE_CASE : Any = pickle.load(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Tuple = tokenizer_new.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : Optional[Any] = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __lowerCAmelCase ( self ) ->str: try: SCREAMING_SNAKE_CASE : Any = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __lowerCAmelCase ( self ) ->str: try: SCREAMING_SNAKE_CASE : str = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : Union[str, Any] = MecabTokenizer(do_lower_case=_lowerCamelCase , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __lowerCAmelCase ( self ) ->Any: try: SCREAMING_SNAKE_CASE : List[str] = MecabTokenizer( do_lower_case=_lowerCamelCase , normalize_text=_lowerCamelCase , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __lowerCAmelCase ( self ) ->Any: SCREAMING_SNAKE_CASE : Any = MecabTokenizer(normalize_text=_lowerCamelCase , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __lowerCAmelCase ( self ) ->str: SCREAMING_SNAKE_CASE : Dict = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[str] = '''こんにちは、世界。\nこんばんは、世界。''' SCREAMING_SNAKE_CASE : Dict = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_lowerCamelCase , '''wb''' ) as handle: pickle.dump(_lowerCamelCase , _lowerCamelCase ) with open(_lowerCamelCase , '''rb''' ) as handle: SCREAMING_SNAKE_CASE : int = pickle.load(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[int] = tokenizer_new.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) @require_sudachi def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : List[Any] = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __lowerCAmelCase ( self ) ->str: SCREAMING_SNAKE_CASE : Any = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __lowerCAmelCase ( self ) ->int: SCREAMING_SNAKE_CASE : Dict = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : str = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __lowerCAmelCase ( self ) ->Dict: SCREAMING_SNAKE_CASE : Dict = SudachiTokenizer(do_lower_case=_lowerCamelCase , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Optional[int] = SudachiTokenizer(normalize_text=_lowerCamelCase , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __lowerCAmelCase ( self ) ->str: SCREAMING_SNAKE_CASE : Any = SudachiTokenizer(trim_whitespace=_lowerCamelCase , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __lowerCAmelCase ( self ) ->str: SCREAMING_SNAKE_CASE : Any = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Dict = '''こんにちは、世界。\nこんばんは、世界。''' SCREAMING_SNAKE_CASE : Any = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_lowerCamelCase , '''wb''' ) as handle: pickle.dump(_lowerCamelCase , _lowerCamelCase ) with open(_lowerCamelCase , '''rb''' ) as handle: SCREAMING_SNAKE_CASE : Any = pickle.load(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Tuple = tokenizer_new.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) @require_jumanpp def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : List[str] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : Dict = JumanppTokenizer(do_lower_case=_lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __lowerCAmelCase ( self ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : Any = JumanppTokenizer(normalize_text=_lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : Optional[int] = JumanppTokenizer(trim_whitespace=_lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __lowerCAmelCase ( self ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : Optional[int] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : Dict = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] SCREAMING_SNAKE_CASE : List[Any] = {} for i, token in enumerate(_lowerCamelCase ): SCREAMING_SNAKE_CASE : Any = i SCREAMING_SNAKE_CASE : Tuple = WordpieceTokenizer(vocab=_lowerCamelCase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __lowerCAmelCase ( self ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : str = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) SCREAMING_SNAKE_CASE : List[str] = tokenizer.subword_tokenizer SCREAMING_SNAKE_CASE : str = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(_lowerCamelCase , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) SCREAMING_SNAKE_CASE : int = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(_lowerCamelCase , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __lowerCAmelCase ( self ) ->Dict: SCREAMING_SNAKE_CASE : Union[str, Any] = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.encode('''ありがとう。''' , add_special_tokens=_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.encode('''どういたしまして。''' , add_special_tokens=_lowerCamelCase ) SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase , _lowerCamelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class a_ ( a__ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = BertJapaneseTokenizer __SCREAMING_SNAKE_CASE : int = False def __lowerCAmelCase ( self ) ->Union[str, Any]: super().setUp() SCREAMING_SNAKE_CASE : Optional[Any] = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __lowerCAmelCase ( self , **_lowerCamelCase ) ->List[Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **_lowerCamelCase ) def __lowerCAmelCase ( self , _lowerCamelCase ) ->List[Any]: SCREAMING_SNAKE_CASE : int = '''こんにちは、世界。 \nこんばんは、世界。''' SCREAMING_SNAKE_CASE : Optional[int] = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __lowerCAmelCase ( self ) ->Optional[int]: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Any: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Union[str, Any]: pass # TODO add if relevant def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Optional[Any] = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) SCREAMING_SNAKE_CASE : List[str] = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( _lowerCamelCase , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __lowerCAmelCase ( self ) ->str: SCREAMING_SNAKE_CASE : Any = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] SCREAMING_SNAKE_CASE : str = {} for i, token in enumerate(_lowerCamelCase ): SCREAMING_SNAKE_CASE : Optional[int] = i SCREAMING_SNAKE_CASE : Dict = CharacterTokenizer(vocab=_lowerCamelCase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : str = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.encode('''ありがとう。''' , add_special_tokens=_lowerCamelCase ) SCREAMING_SNAKE_CASE : str = tokenizer.encode('''どういたしまして。''' , add_special_tokens=_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase , _lowerCamelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class a_ ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : str = '''cl-tohoku/bert-base-japanese''' SCREAMING_SNAKE_CASE : Tuple = AutoTokenizer.from_pretrained(_lowerCamelCase ) self.assertIsInstance(_lowerCamelCase , _lowerCamelCase ) class a_ ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : Union[str, Any] = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(_lowerCamelCase ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) SCREAMING_SNAKE_CASE : Optional[Any] = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(_lowerCamelCase ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
19
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig a__ : List[str] = logging.get_logger(__name__) # General docstring a__ : Tuple = '''MobileNetV1Config''' # Base docstring a__ : Optional[Any] = '''google/mobilenet_v1_1.0_224''' a__ : Tuple = [1, 1_024, 7, 7] # Image classification docstring a__ : Optional[int] = '''google/mobilenet_v1_1.0_224''' a__ : int = '''tabby, tabby cat''' a__ : List[Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def UpperCAmelCase_( a__ , a__ , a__=None ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = {} if isinstance(a__ , a__ ): SCREAMING_SNAKE_CASE : List[str] = model.mobilenet_va else: SCREAMING_SNAKE_CASE : Union[str, Any] = model SCREAMING_SNAKE_CASE : Optional[int] = '''MobilenetV1/Conv2d_0/''' SCREAMING_SNAKE_CASE : Tuple = backbone.conv_stem.convolution.weight SCREAMING_SNAKE_CASE : Tuple = backbone.conv_stem.normalization.bias SCREAMING_SNAKE_CASE : Optional[Any] = backbone.conv_stem.normalization.weight SCREAMING_SNAKE_CASE : Union[str, Any] = backbone.conv_stem.normalization.running_mean SCREAMING_SNAKE_CASE : Any = backbone.conv_stem.normalization.running_var for i in range(13 ): SCREAMING_SNAKE_CASE : Dict = i + 1 SCREAMING_SNAKE_CASE : Union[str, Any] = i * 2 SCREAMING_SNAKE_CASE : Any = backbone.layer[pt_index] SCREAMING_SNAKE_CASE : Optional[Any] = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" SCREAMING_SNAKE_CASE : Any = pointer.convolution.weight SCREAMING_SNAKE_CASE : Tuple = pointer.normalization.bias SCREAMING_SNAKE_CASE : List[Any] = pointer.normalization.weight SCREAMING_SNAKE_CASE : Optional[Any] = pointer.normalization.running_mean SCREAMING_SNAKE_CASE : List[Any] = pointer.normalization.running_var SCREAMING_SNAKE_CASE : List[Any] = backbone.layer[pt_index + 1] SCREAMING_SNAKE_CASE : Any = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" SCREAMING_SNAKE_CASE : Dict = pointer.convolution.weight SCREAMING_SNAKE_CASE : Optional[Any] = pointer.normalization.bias SCREAMING_SNAKE_CASE : Optional[Any] = pointer.normalization.weight SCREAMING_SNAKE_CASE : int = pointer.normalization.running_mean SCREAMING_SNAKE_CASE : str = pointer.normalization.running_var if isinstance(a__ , a__ ): SCREAMING_SNAKE_CASE : List[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' SCREAMING_SNAKE_CASE : List[str] = model.classifier.weight SCREAMING_SNAKE_CASE : List[str] = model.classifier.bias return tf_to_pt_map def UpperCAmelCase_( a__ , a__ , a__ ): """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model SCREAMING_SNAKE_CASE : Optional[Any] = tf.train.list_variables(a__ ) SCREAMING_SNAKE_CASE : List[Any] = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) SCREAMING_SNAKE_CASE : Tuple = tf.train.load_variable(a__ , a__ ) SCREAMING_SNAKE_CASE : Dict = array # Build TF to PyTorch weights loading map SCREAMING_SNAKE_CASE : int = _build_tf_to_pytorch_map(a__ , a__ , a__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue SCREAMING_SNAKE_CASE : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) SCREAMING_SNAKE_CASE : Tuple = np.transpose(a__ , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer SCREAMING_SNAKE_CASE : Union[str, Any] = array.squeeze().transpose() else: SCREAMING_SNAKE_CASE : Optional[int] = np.transpose(a__ , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) SCREAMING_SNAKE_CASE : Tuple = torch.from_numpy(a__ ) tf_weights.pop(a__ , a__ ) tf_weights.pop(name + '''/RMSProp''' , a__ ) tf_weights.pop(name + '''/RMSProp_1''' , a__ ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , a__ ) logger.info(F"""Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}""" ) return model def UpperCAmelCase_( a__ , a__ ): """simple docstring""" SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = features.shape[-2:] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = conv_layer.stride SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = conv_layer.kernel_size if in_height % stride_height == 0: SCREAMING_SNAKE_CASE : List[str] = max(kernel_height - stride_height , 0 ) else: SCREAMING_SNAKE_CASE : str = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: SCREAMING_SNAKE_CASE : int = max(kernel_width - stride_width , 0 ) else: SCREAMING_SNAKE_CASE : Tuple = max(kernel_width - (in_width % stride_width) , 0 ) SCREAMING_SNAKE_CASE : List[str] = pad_along_width // 2 SCREAMING_SNAKE_CASE : Any = pad_along_width - pad_left SCREAMING_SNAKE_CASE : str = pad_along_height // 2 SCREAMING_SNAKE_CASE : Optional[int] = pad_along_height - pad_top SCREAMING_SNAKE_CASE : List[Any] = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(a__ , a__ , '''constant''' , 0.0 ) class a_ ( nn.Module ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 1 , _lowerCamelCase = 1 , _lowerCamelCase = False , _lowerCamelCase = True , _lowerCamelCase = True , ) ->None: super().__init__() SCREAMING_SNAKE_CASE : Any = config if in_channels % groups != 0: raise ValueError(F"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(F"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) SCREAMING_SNAKE_CASE : Any = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) SCREAMING_SNAKE_CASE : List[str] = nn.Convad( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , kernel_size=_lowerCamelCase , stride=_lowerCamelCase , padding=_lowerCamelCase , groups=_lowerCamelCase , bias=_lowerCamelCase , padding_mode='''zeros''' , ) if use_normalization: SCREAMING_SNAKE_CASE : List[Any] = nn.BatchNormad( num_features=_lowerCamelCase , eps=config.layer_norm_eps , momentum=0.9_9_9_7 , affine=_lowerCamelCase , track_running_stats=_lowerCamelCase , ) else: SCREAMING_SNAKE_CASE : Dict = None if use_activation: if isinstance(_lowerCamelCase , _lowerCamelCase ): SCREAMING_SNAKE_CASE : Any = ACTaFN[use_activation] elif isinstance(config.hidden_act , _lowerCamelCase ): SCREAMING_SNAKE_CASE : List[str] = ACTaFN[config.hidden_act] else: SCREAMING_SNAKE_CASE : List[Any] = config.hidden_act else: SCREAMING_SNAKE_CASE : Optional[Any] = None def __lowerCAmelCase ( self , _lowerCamelCase ) ->torch.Tensor: if self.config.tf_padding: SCREAMING_SNAKE_CASE : List[Any] = apply_tf_padding(_lowerCamelCase , self.convolution ) SCREAMING_SNAKE_CASE : Dict = self.convolution(_lowerCamelCase ) if self.normalization is not None: SCREAMING_SNAKE_CASE : int = self.normalization(_lowerCamelCase ) if self.activation is not None: SCREAMING_SNAKE_CASE : List[Any] = self.activation(_lowerCamelCase ) return features class a_ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = MobileNetVaConfig __SCREAMING_SNAKE_CASE : List[Any] = load_tf_weights_in_mobilenet_va __SCREAMING_SNAKE_CASE : int = 'mobilenet_v1' __SCREAMING_SNAKE_CASE : int = 'pixel_values' __SCREAMING_SNAKE_CASE : List[str] = False def __lowerCAmelCase ( self , _lowerCamelCase ) ->None: if isinstance(_lowerCamelCase , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(_lowerCamelCase , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) a__ : str = r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' a__ : Union[str, Any] = r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( 'The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.' , a__ , ) class a_ ( a__ ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True ) ->Dict: super().__init__(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[Any] = config SCREAMING_SNAKE_CASE : Dict = 32 SCREAMING_SNAKE_CASE : Optional[Any] = max(int(depth * config.depth_multiplier ) , config.min_depth ) SCREAMING_SNAKE_CASE : str = MobileNetVaConvLayer( _lowerCamelCase , in_channels=config.num_channels , out_channels=_lowerCamelCase , kernel_size=3 , stride=2 , ) SCREAMING_SNAKE_CASE : Union[str, Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] SCREAMING_SNAKE_CASE : Any = nn.ModuleList() for i in range(13 ): SCREAMING_SNAKE_CASE : int = out_channels if strides[i] == 2 or i == 0: depth *= 2 SCREAMING_SNAKE_CASE : Tuple = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( _lowerCamelCase , in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , kernel_size=3 , stride=strides[i] , groups=_lowerCamelCase , ) ) self.layer.append( MobileNetVaConvLayer( _lowerCamelCase , in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , kernel_size=1 , ) ) SCREAMING_SNAKE_CASE : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def __lowerCAmelCase ( self , _lowerCamelCase ) ->List[str]: raise NotImplementedError @add_start_docstrings_to_model_forward(_lowerCamelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_lowerCamelCase , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __lowerCAmelCase ( self , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , ) ->Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: SCREAMING_SNAKE_CASE : Union[str, Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) SCREAMING_SNAKE_CASE : List[Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) SCREAMING_SNAKE_CASE : Union[str, Any] = self.conv_stem(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Tuple = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): SCREAMING_SNAKE_CASE : Optional[int] = layer_module(_lowerCamelCase ) if output_hidden_states: SCREAMING_SNAKE_CASE : List[str] = all_hidden_states + (hidden_states,) SCREAMING_SNAKE_CASE : List[str] = hidden_states if self.pooler is not None: SCREAMING_SNAKE_CASE : Tuple = torch.flatten(self.pooler(_lowerCamelCase ) , start_dim=1 ) else: SCREAMING_SNAKE_CASE : List[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_lowerCamelCase , pooler_output=_lowerCamelCase , hidden_states=_lowerCamelCase , ) @add_start_docstrings( '\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ' , a__ , ) class a_ ( a__ ): """simple docstring""" def __init__( self , _lowerCamelCase ) ->None: super().__init__(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[int] = config.num_labels SCREAMING_SNAKE_CASE : str = MobileNetVaModel(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Dict = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head SCREAMING_SNAKE_CASE : Optional[int] = nn.Dropout(config.classifier_dropout_prob , inplace=_lowerCamelCase ) SCREAMING_SNAKE_CASE : int = nn.Linear(_lowerCamelCase , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_lowerCamelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_lowerCamelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __lowerCAmelCase ( self , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , ) ->Union[tuple, ImageClassifierOutputWithNoAttention]: SCREAMING_SNAKE_CASE : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict SCREAMING_SNAKE_CASE : Dict = self.mobilenet_va(_lowerCamelCase , output_hidden_states=_lowerCamelCase , return_dict=_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = outputs.pooler_output if return_dict else outputs[1] SCREAMING_SNAKE_CASE : Tuple = self.classifier(self.dropout(_lowerCamelCase ) ) SCREAMING_SNAKE_CASE : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: SCREAMING_SNAKE_CASE : Any = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): SCREAMING_SNAKE_CASE : Optional[int] = '''single_label_classification''' else: SCREAMING_SNAKE_CASE : Dict = '''multi_label_classification''' if self.config.problem_type == "regression": SCREAMING_SNAKE_CASE : Any = MSELoss() if self.num_labels == 1: SCREAMING_SNAKE_CASE : List[Any] = loss_fct(logits.squeeze() , labels.squeeze() ) else: SCREAMING_SNAKE_CASE : Dict = loss_fct(_lowerCamelCase , _lowerCamelCase ) elif self.config.problem_type == "single_label_classification": SCREAMING_SNAKE_CASE : str = CrossEntropyLoss() SCREAMING_SNAKE_CASE : int = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": SCREAMING_SNAKE_CASE : List[Any] = BCEWithLogitsLoss() SCREAMING_SNAKE_CASE : List[Any] = loss_fct(_lowerCamelCase , _lowerCamelCase ) if not return_dict: SCREAMING_SNAKE_CASE : Optional[Any] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=_lowerCamelCase , logits=_lowerCamelCase , hidden_states=outputs.hidden_states , )
19
1
from ...processing_utils import ProcessorMixin class lowercase_ ( UpperCAmelCase__ ): A__ : str = """SpeechT5FeatureExtractor""" A__ : Dict = """SpeechT5Tokenizer""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" UpperCamelCase_ = kwargs.pop("""audio""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""text""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""text_target""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""audio_target""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""sampling_rate""" , UpperCAmelCase__ ) if audio is not None and text is not None: raise ValueError( """Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?""" ) if audio_target is not None and text_target is not None: raise ValueError( """Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?""" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( """You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.""" ) if audio is not None: UpperCamelCase_ = self.feature_extractor(UpperCAmelCase__ , *UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , **UpperCAmelCase__ ) elif text is not None: UpperCamelCase_ = self.tokenizer(UpperCAmelCase__ , **UpperCAmelCase__ ) else: UpperCamelCase_ = None if audio_target is not None: UpperCamelCase_ = self.feature_extractor(audio_target=UpperCAmelCase__ , *UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , **UpperCAmelCase__ ) UpperCamelCase_ = targets["""input_values"""] elif text_target is not None: UpperCamelCase_ = self.tokenizer(UpperCAmelCase__ , **UpperCAmelCase__ ) UpperCamelCase_ = targets["""input_ids"""] else: UpperCamelCase_ = None if inputs is None: return targets if targets is not None: UpperCamelCase_ = labels UpperCamelCase_ = targets.get("""attention_mask""" ) if decoder_attention_mask is not None: UpperCamelCase_ = decoder_attention_mask return inputs def lowerCamelCase_ ( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" UpperCamelCase_ = kwargs.pop("""input_values""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""input_ids""" , UpperCAmelCase__ ) UpperCamelCase_ = kwargs.pop("""labels""" , UpperCAmelCase__ ) if input_values is not None and input_ids is not None: raise ValueError("""Cannot process both `input_values` and `input_ids` inputs.""" ) if input_values is None and input_ids is None and labels is None: raise ValueError( """You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.""" ) if input_values is not None: UpperCamelCase_ = self.feature_extractor.pad(UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) elif input_ids is not None: UpperCamelCase_ = self.tokenizer.pad(UpperCAmelCase__ , **UpperCAmelCase__ ) else: UpperCamelCase_ = None if labels is not None: if "input_ids" in labels or (isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) and "input_ids" in labels[0]): UpperCamelCase_ = self.tokenizer.pad(UpperCAmelCase__ , **UpperCAmelCase__ ) UpperCamelCase_ = targets["""input_ids"""] else: UpperCamelCase_ = self.feature_extractor.feature_size UpperCamelCase_ = self.feature_extractor.num_mel_bins UpperCamelCase_ = self.feature_extractor.pad(UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) UpperCamelCase_ = feature_size_hack UpperCamelCase_ = targets["""input_values"""] else: UpperCamelCase_ = None if inputs is None: return targets if targets is not None: UpperCamelCase_ = labels UpperCamelCase_ = targets.get("""attention_mask""" ) if decoder_attention_mask is not None: UpperCamelCase_ = decoder_attention_mask return inputs def lowerCamelCase_ ( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCamelCase_ ( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ )
122
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Dict=3 , UpperCAmelCase__ : List[Any]=30 , UpperCAmelCase__ : Any=400 , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Optional[Any]=[0.5, 0.5, 0.5] , UpperCAmelCase__ : Any=[0.5, 0.5, 0.5] , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : Optional[int]=1 / 255 , UpperCAmelCase__ : Optional[Any]=True , ) ->str: '''simple docstring''' A__ = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1_333} A__ = parent A__ = batch_size A__ = num_channels A__ = min_resolution A__ = max_resolution A__ = do_resize A__ = size A__ = do_normalize A__ = image_mean A__ = image_std A__ = do_rescale A__ = rescale_factor A__ = do_pad def SCREAMING_SNAKE_CASE ( self : Any) ->List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int=False) ->Optional[Any]: '''simple docstring''' if not batched: A__ = image_inputs[0] if isinstance(UpperCAmelCase__ , Image.Image): A__ , A__ = image.size else: A__ , A__ = image.shape[1], image.shape[2] if w < h: A__ = int(self.size['''shortest_edge'''] * h / w) A__ = self.size['''shortest_edge'''] elif w > h: A__ = self.size['''shortest_edge'''] A__ = int(self.size['''shortest_edge'''] * w / h) else: A__ = self.size['''shortest_edge'''] A__ = self.size['''shortest_edge'''] else: A__ = [] for image in image_inputs: A__ , A__ = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) A__ = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__: item[0])[0] A__ = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__: item[1])[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = DeformableDetrImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' A__ = DeformableDetrImageProcessingTester(self) @property def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Any: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : List[str]) ->List[str]: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCAmelCase__ , '''image_mean''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''image_std''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_normalize''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_rescale''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_pad''')) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''')) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->int: '''simple docstring''' A__ = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1_333}) self.assertEqual(image_processor.do_pad , UpperCAmelCase__) A__ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=UpperCAmelCase__) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84}) self.assertEqual(image_processor.do_pad , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Any) ->List[str]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Dict: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict) # create random PIL images A__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image) # Test not batched input A__ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__) A__ = image_processing(UpperCAmelCase__ , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE ( self : int) ->Optional[int]: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors A__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray) # Test not batched input A__ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A__ = image_processing(UpperCAmelCase__ , return_tensors='''pt''').pixel_values A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE ( self : int) ->Tuple: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors A__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor) # Test not batched input A__ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A__ = image_processing(UpperCAmelCase__ , return_tensors='''pt''').pixel_values A__ , A__ = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->List[str]: '''simple docstring''' A__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''') as f: A__ = json.loads(f.read()) A__ = {'''image_id''': 39_769, '''annotations''': target} # encode them A__ = DeformableDetrImageProcessor() A__ = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , return_tensors='''pt''') # verify pixel values A__ = torch.Size([1, 3, 800, 1_066]) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase__) A__ = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase__ , atol=1e-4)) # verify area A__ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase__)) # verify boxes A__ = torch.Size([6, 4]) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase__) A__ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase__ , atol=1e-3)) # verify image_id A__ = torch.tensor([39_769]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase__)) # verify is_crowd A__ = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase__)) # verify class_labels A__ = torch.tensor([75, 75, 63, 65, 17, 17]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase__)) # verify orig_size A__ = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase__)) # verify size A__ = torch.tensor([800, 1_066]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase__)) @slow def SCREAMING_SNAKE_CASE ( self : Dict) ->Optional[int]: '''simple docstring''' A__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''') as f: A__ = json.loads(f.read()) A__ = {'''file_name''': '''000000039769.png''', '''image_id''': 39_769, '''segments_info''': target} A__ = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''') # encode them A__ = DeformableDetrImageProcessor(format='''coco_panoptic''') A__ = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , masks_path=UpperCAmelCase__ , return_tensors='''pt''') # verify pixel values A__ = torch.Size([1, 3, 800, 1_066]) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase__) A__ = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase__ , atol=1e-4)) # verify area A__ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase__)) # verify boxes A__ = torch.Size([6, 4]) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase__) A__ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase__ , atol=1e-3)) # verify image_id A__ = torch.tensor([39_769]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase__)) # verify is_crowd A__ = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase__)) # verify class_labels A__ = torch.tensor([17, 17, 63, 75, 75, 93]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase__)) # verify masks A__ = 822_873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , UpperCAmelCase__) # verify orig_size A__ = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase__)) # verify size A__ = torch.tensor([800, 1_066]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase__))
14
0
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _A ( ): """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(lowerCAmelCase_ ): requests.request("GET" , "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET" , "https://huggingface.co" , timeout=1.0 ) @pytest.mark.integration def _A ( ): """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET" , "https://huggingface.co" ) def _A ( ): """simple docstring""" with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(lowerCAmelCase_ ): http_head("https://huggingface.co" )
221
import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# UpperCamelCase = [ # (stable-diffusion, HF Diffusers) ('time_embed.0.weight', 'time_embedding.linear_1.weight'), ('time_embed.0.bias', 'time_embedding.linear_1.bias'), ('time_embed.2.weight', 'time_embedding.linear_2.weight'), ('time_embed.2.bias', 'time_embedding.linear_2.bias'), ('input_blocks.0.0.weight', 'conv_in.weight'), ('input_blocks.0.0.bias', 'conv_in.bias'), ('out.0.weight', 'conv_norm_out.weight'), ('out.0.bias', 'conv_norm_out.bias'), ('out.2.weight', 'conv_out.weight'), ('out.2.bias', 'conv_out.bias'), ] UpperCamelCase = [ # (stable-diffusion, HF Diffusers) ('in_layers.0', 'norm1'), ('in_layers.2', 'conv1'), ('out_layers.0', 'norm2'), ('out_layers.3', 'conv2'), ('emb_layers.1', 'time_emb_proj'), ('skip_connection', 'conv_shortcut'), ] UpperCamelCase = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks UpperCamelCase = F"""down_blocks.{i}.resnets.{j}.""" UpperCamelCase = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 UpperCamelCase = F"""down_blocks.{i}.attentions.{j}.""" UpperCamelCase = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks UpperCamelCase = F"""up_blocks.{i}.resnets.{j}.""" UpperCamelCase = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 UpperCamelCase = F"""up_blocks.{i}.attentions.{j}.""" UpperCamelCase = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 UpperCamelCase = F"""down_blocks.{i}.downsamplers.0.conv.""" UpperCamelCase = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 UpperCamelCase = F"""up_blocks.{i}.upsamplers.0.""" UpperCamelCase = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) UpperCamelCase = 'mid_block.attentions.0.' UpperCamelCase = 'middle_block.1.' unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): UpperCamelCase = F"""mid_block.resnets.{j}.""" UpperCamelCase = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def _A ( lowerCAmelCase_ : Any ): """simple docstring""" lowerCAmelCase__ = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: lowerCAmelCase__ = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: lowerCAmelCase__ = v.replace(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase__ = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: lowerCAmelCase__ = v.replace(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase__ = v lowerCAmelCase__ = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# UpperCamelCase = [ # (stable-diffusion, HF Diffusers) ('nin_shortcut', 'conv_shortcut'), ('norm_out', 'conv_norm_out'), ('mid.attn_1.', 'mid_block.attentions.0.'), ] for i in range(4): # down_blocks have two resnets for j in range(2): UpperCamelCase = F"""encoder.down_blocks.{i}.resnets.{j}.""" UpperCamelCase = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: UpperCamelCase = F"""down_blocks.{i}.downsamplers.0.""" UpperCamelCase = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) UpperCamelCase = F"""up_blocks.{i}.upsamplers.0.""" UpperCamelCase = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): UpperCamelCase = F"""decoder.up_blocks.{i}.resnets.{j}.""" UpperCamelCase = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): UpperCamelCase = F"""mid_block.resnets.{i}.""" UpperCamelCase = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) UpperCamelCase = [ # (stable-diffusion, HF Diffusers) ('norm.', 'group_norm.'), ('q.', 'query.'), ('k.', 'key.'), ('v.', 'value.'), ('proj_out.', 'proj_attn.'), ] def _A ( lowerCAmelCase_ : List[str] ): """simple docstring""" return w.reshape(*w.shape , 1 , 1 ) def _A ( lowerCAmelCase_ : Dict ): """simple docstring""" lowerCAmelCase__ = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: lowerCAmelCase__ = v.replace(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase__ = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: lowerCAmelCase__ = v.replace(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase__ = v lowerCAmelCase__ = {v: vae_state_dict[k] for k, v in mapping.items()} lowerCAmelCase__ = ["q", "k", "v", "proj_out"] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'mid.attn_1.{weight_name}.weight' in k: print(F'Reshaping {k} for SD format' ) lowerCAmelCase__ = reshape_weight_for_sd(lowerCAmelCase_ ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# UpperCamelCase = [ # (stable-diffusion, HF Diffusers) ('resblocks.', 'text_model.encoder.layers.'), ('ln_1', 'layer_norm1'), ('ln_2', 'layer_norm2'), ('.c_fc.', '.fc1.'), ('.c_proj.', '.fc2.'), ('.attn', '.self_attn'), ('ln_final.', 'transformer.text_model.final_layer_norm.'), ('token_embedding.weight', 'transformer.text_model.embeddings.token_embedding.weight'), ('positional_embedding', 'transformer.text_model.embeddings.position_embedding.weight'), ] UpperCamelCase = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} UpperCamelCase = re.compile('|'.join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp UpperCamelCase = {'q': 0, 'k': 1, 'v': 2} def _A ( lowerCAmelCase_ : Dict ): """simple docstring""" lowerCAmelCase__ = {} lowerCAmelCase__ = {} lowerCAmelCase__ = {} for k, v in text_enc_dict.items(): if ( k.endswith(".self_attn.q_proj.weight" ) or k.endswith(".self_attn.k_proj.weight" ) or k.endswith(".self_attn.v_proj.weight" ) ): lowerCAmelCase__ = k[: -len(".q_proj.weight" )] lowerCAmelCase__ = k[-len("q_proj.weight" )] if k_pre not in capture_qkv_weight: lowerCAmelCase__ = [None, None, None] lowerCAmelCase__ = v continue if ( k.endswith(".self_attn.q_proj.bias" ) or k.endswith(".self_attn.k_proj.bias" ) or k.endswith(".self_attn.v_proj.bias" ) ): lowerCAmelCase__ = k[: -len(".q_proj.bias" )] lowerCAmelCase__ = k[-len("q_proj.bias" )] if k_pre not in capture_qkv_bias: lowerCAmelCase__ = [None, None, None] lowerCAmelCase__ = v continue lowerCAmelCase__ = textenc_pattern.sub(lambda lowerCAmelCase_ : protected[re.escape(m.group(0 ) )] , lowerCAmelCase_ ) lowerCAmelCase__ = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception("CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing" ) lowerCAmelCase__ = textenc_pattern.sub(lambda lowerCAmelCase_ : protected[re.escape(m.group(0 ) )] , lowerCAmelCase_ ) lowerCAmelCase__ = torch.cat(lowerCAmelCase_ ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception("CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing" ) lowerCAmelCase__ = textenc_pattern.sub(lambda lowerCAmelCase_ : protected[re.escape(m.group(0 ) )] , lowerCAmelCase_ ) lowerCAmelCase__ = torch.cat(lowerCAmelCase_ ) return new_state_dict def _A ( lowerCAmelCase_ : Any ): """simple docstring""" return text_enc_dict if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--half', action='store_true', help='Save weights in half precision.') parser.add_argument( '--use_safetensors', action='store_true', help='Save weights use safetensors, default is ckpt.' ) UpperCamelCase = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors UpperCamelCase = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.safetensors') UpperCamelCase = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.safetensors') UpperCamelCase = osp.join(args.model_path, 'text_encoder', 'model.safetensors') # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): UpperCamelCase = load_file(unet_path, device='cpu') else: UpperCamelCase = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.bin') UpperCamelCase = torch.load(unet_path, map_location='cpu') if osp.exists(vae_path): UpperCamelCase = load_file(vae_path, device='cpu') else: UpperCamelCase = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.bin') UpperCamelCase = torch.load(vae_path, map_location='cpu') if osp.exists(text_enc_path): UpperCamelCase = load_file(text_enc_path, device='cpu') else: UpperCamelCase = osp.join(args.model_path, 'text_encoder', 'pytorch_model.bin') UpperCamelCase = torch.load(text_enc_path, map_location='cpu') # Convert the UNet model UpperCamelCase = convert_unet_state_dict(unet_state_dict) UpperCamelCase = {'model.diffusion_model.' + k: v for k, v in unet_state_dict.items()} # Convert the VAE model UpperCamelCase = convert_vae_state_dict(vae_state_dict) UpperCamelCase = {'first_stage_model.' + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper UpperCamelCase = 'text_model.encoder.layers.22.layer_norm2.bias' in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm UpperCamelCase = {'transformer.' + k: v for k, v in text_enc_dict.items()} UpperCamelCase = convert_text_enc_state_dict_vaa(text_enc_dict) UpperCamelCase = {'cond_stage_model.model.' + k: v for k, v in text_enc_dict.items()} else: UpperCamelCase = convert_text_enc_state_dict(text_enc_dict) UpperCamelCase = {'cond_stage_model.transformer.' + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint UpperCamelCase = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: UpperCamelCase = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: UpperCamelCase = {'state_dict': state_dict} torch.save(state_dict, args.checkpoint_path)
221
1
"""simple docstring""" from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar __SCREAMING_SNAKE_CASE : List[str] = TypeVar('T') class __A (Generic[T]): '''simple docstring''' __lowercase: Tuple = 42 # Cache store of keys __lowercase: int = 42 # References of the keys in cache __lowercase: Dict = 10 # Maximum capacity of cache def __init__( self : Optional[Any] , UpperCAmelCase_ : Tuple ) ->None: """simple docstring""" snake_case_ = deque() snake_case_ = set() if not n: snake_case_ = sys.maxsize elif n < 0: raise ValueError("""n should be an integer greater than 0.""" ) else: snake_case_ = n def lowerCAmelCase ( self : Union[str, Any] , UpperCAmelCase_ : Any ) ->None: """simple docstring""" if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: snake_case_ = self.dq_store.pop() self.key_reference.remove(UpperCAmelCase_ ) else: self.dq_store.remove(UpperCAmelCase_ ) self.dq_store.appendleft(UpperCAmelCase_ ) self.key_reference.add(UpperCAmelCase_ ) def lowerCAmelCase ( self : List[Any] ) ->None: """simple docstring""" for k in self.dq_store: print(UpperCAmelCase_ ) def __repr__( self : int ) ->str: """simple docstring""" return F"""LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}""" if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : LRUCache[str | int] = LRUCache(4) lru_cache.refer('A') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('A') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
347
'''simple docstring''' from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def __lowerCamelCase ( _lowercase ) -> Optional[Any]: return getitem, k def __lowerCamelCase ( _lowercase , _lowercase ) -> List[str]: return setitem, k, v def __lowerCamelCase ( _lowercase ) -> int: return delitem, k def __lowerCamelCase ( _lowercase , _lowercase , *_lowercase ) -> Optional[Any]: try: return fun(_lowercase , *_lowercase ), None except Exception as e: return None, e a : List[str] = ( _set("""key_a""", """val_a"""), _set("""key_b""", """val_b"""), ) a : List[Any] = [ _set("""key_a""", """val_a"""), _set("""key_a""", """val_b"""), ] a : int = [ _set("""key_a""", """val_a"""), _set("""key_b""", """val_b"""), _del("""key_a"""), _del("""key_b"""), _set("""key_a""", """val_a"""), _del("""key_a"""), ] a : List[Any] = [ _get("""key_a"""), _del("""key_a"""), _set("""key_a""", """val_a"""), _del("""key_a"""), _del("""key_a"""), _get("""key_a"""), ] a : Tuple = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] a : Optional[Any] = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set("""key_a""", """val_b"""), ] @pytest.mark.parametrize( """operations""" , ( pytest.param(_add_items , id="""add items""" ), pytest.param(_overwrite_items , id="""overwrite items""" ), pytest.param(_delete_items , id="""delete items""" ), pytest.param(_access_absent_items , id="""access absent items""" ), pytest.param(_add_with_resize_up , id="""add with resize up""" ), pytest.param(_add_with_resize_down , id="""add with resize down""" ), ) , ) def __lowerCamelCase ( _lowercase ) -> Optional[int]: UpperCAmelCase : List[str] = HashMap(initial_block_size=4 ) UpperCAmelCase : Dict = {} for _, (fun, *args) in enumerate(_lowercase ): UpperCAmelCase , UpperCAmelCase : Union[str, Any] = _run_operation(_lowercase , _lowercase , *_lowercase ) UpperCAmelCase , UpperCAmelCase : Any = _run_operation(_lowercase , _lowercase , *_lowercase ) assert my_res == py_res assert str(_lowercase ) == str(_lowercase ) assert set(_lowercase ) == set(_lowercase ) assert len(_lowercase ) == len(_lowercase ) assert set(my.items() ) == set(py.items() ) def __lowerCamelCase ( ) -> List[Any]: def is_public(_lowercase ) -> bool: return not name.startswith("""_""" ) UpperCAmelCase : int = {name for name in dir({} ) if is_public(_lowercase )} UpperCAmelCase : Any = {name for name in dir(HashMap() ) if is_public(_lowercase )} assert dict_public_names > hash_public_names
265
0
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=__lowerCamelCase ) class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = field(default="""question-answering-extractive""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowerCAmelCase__ = Features({"""question""": Value("""string""" ), """context""": Value("""string""" )} ) lowerCAmelCase__ = Features( { """answers""": Sequence( { """text""": Value("""string""" ), """answer_start""": Value("""int32""" ), } ) } ) lowerCAmelCase__ = """question""" lowerCAmelCase__ = """context""" lowerCAmelCase__ = """answers""" @property def UpperCAmelCase__ ( self : Optional[Any] ): return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
360
import math def A__ ( SCREAMING_SNAKE_CASE__) -> int: if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): __snake_case: Optional[int] = F'''Input value of [number={number}] must be an integer''' raise TypeError(SCREAMING_SNAKE_CASE__) if number < 1: __snake_case: Optional[int] = F'''Input value of [number={number}] must be > 0''' raise ValueError(SCREAMING_SNAKE_CASE__) elif number == 1: return 3 elif number == 2: return 5 else: __snake_case: List[Any] = int(math.log(number // 3 , 2)) + 2 __snake_case: str = [3, 5] __snake_case: int = 2 __snake_case: List[str] = 3 for block in range(1 , SCREAMING_SNAKE_CASE__): for _ in range(SCREAMING_SNAKE_CASE__): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1]) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): __UpperCAmelCase : Optional[int] = 0 try: __UpperCAmelCase : int = proth(number) except ValueError: print(f'ValueError: there is no {number}th Proth number') continue print(f'The {number}th Proth number: {value}')
293
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class _a ( _lowerCAmelCase ): A = '''mvp''' A = ['''past_key_values'''] A = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__(self, SCREAMING_SNAKE_CASE_=50267, SCREAMING_SNAKE_CASE_=1024, SCREAMING_SNAKE_CASE_=12, SCREAMING_SNAKE_CASE_=4096, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=12, SCREAMING_SNAKE_CASE_=4096, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=1024, SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_=0.0_2, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=1, SCREAMING_SNAKE_CASE_=0, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=100, SCREAMING_SNAKE_CASE_=800, **SCREAMING_SNAKE_CASE_, ) -> Union[str, Any]: UpperCAmelCase_: Tuple = vocab_size UpperCAmelCase_: Union[str, Any] = max_position_embeddings UpperCAmelCase_: Optional[int] = d_model UpperCAmelCase_: List[str] = encoder_ffn_dim UpperCAmelCase_: str = encoder_layers UpperCAmelCase_: str = encoder_attention_heads UpperCAmelCase_: Optional[int] = decoder_ffn_dim UpperCAmelCase_: int = decoder_layers UpperCAmelCase_: Union[str, Any] = decoder_attention_heads UpperCAmelCase_: Optional[Any] = dropout UpperCAmelCase_: Any = attention_dropout UpperCAmelCase_: List[str] = activation_dropout UpperCAmelCase_: Tuple = activation_function UpperCAmelCase_: Optional[Any] = init_std UpperCAmelCase_: int = encoder_layerdrop UpperCAmelCase_: Any = decoder_layerdrop UpperCAmelCase_: str = classifier_dropout UpperCAmelCase_: Optional[Any] = use_cache UpperCAmelCase_: Optional[Any] = encoder_layers UpperCAmelCase_: Any = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase_: List[Any] = use_prompt UpperCAmelCase_: Dict = prompt_length UpperCAmelCase_: str = prompt_mid_dim super().__init__( pad_token_id=SCREAMING_SNAKE_CASE_, bos_token_id=SCREAMING_SNAKE_CASE_, eos_token_id=SCREAMING_SNAKE_CASE_, is_encoder_decoder=SCREAMING_SNAKE_CASE_, decoder_start_token_id=SCREAMING_SNAKE_CASE_, forced_eos_token_id=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_, ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""", SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Dict = self.bos_token_id warnings.warn( f'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' """The config can simply be saved and uploaded again to be fixed.""" )
147
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Any = logging.get_logger(__name__) class _a ( _lowerCAmelCase ): A = ['''pixel_values'''] def __init__(self, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = 1 / 255, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> None: super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: int = size if size is not None else {"""height""": 256, """width""": 256} UpperCAmelCase_: str = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Union[str, Any] = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} UpperCAmelCase_: Any = get_size_dict(SCREAMING_SNAKE_CASE_, param_name="""crop_size""" ) UpperCAmelCase_: Dict = do_resize UpperCAmelCase_: Tuple = size UpperCAmelCase_: Dict = resample UpperCAmelCase_: Union[str, Any] = do_center_crop UpperCAmelCase_: List[str] = crop_size UpperCAmelCase_: Optional[int] = do_rescale UpperCAmelCase_: Dict = rescale_factor UpperCAmelCase_: Optional[Any] = do_normalize UpperCAmelCase_: Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_: Union[str, Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCAmelCase_: int = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( SCREAMING_SNAKE_CASE_, size=(size["""height"""], size["""width"""]), resample=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCAmelCase_: int = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(SCREAMING_SNAKE_CASE_, size=(size["""height"""], size["""width"""]), data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> str: return rescale(SCREAMING_SNAKE_CASE_, scale=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: return normalize(SCREAMING_SNAKE_CASE_, mean=SCREAMING_SNAKE_CASE_, std=SCREAMING_SNAKE_CASE_, data_format=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST, **SCREAMING_SNAKE_CASE_, ) -> PIL.Image.Image: UpperCAmelCase_: str = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_: str = resample if resample is not None else self.resample UpperCAmelCase_: Any = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_: str = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_: List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_: str = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_: Optional[int] = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_: Optional[Any] = image_std if image_std is not None else self.image_std UpperCAmelCase_: List[str] = size if size is not None else self.size UpperCAmelCase_: Any = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[Any] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_: Dict = get_size_dict(SCREAMING_SNAKE_CASE_, param_name="""crop_size""" ) UpperCAmelCase_: Any = make_list_of_images(SCREAMING_SNAKE_CASE_ ) if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase_: List[str] = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] if do_resize: UpperCAmelCase_: Any = [self.resize(image=SCREAMING_SNAKE_CASE_, size=SCREAMING_SNAKE_CASE_, resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_center_crop: UpperCAmelCase_: Dict = [self.center_crop(image=SCREAMING_SNAKE_CASE_, size=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: UpperCAmelCase_: str = [self.rescale(image=SCREAMING_SNAKE_CASE_, scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: UpperCAmelCase_: Optional[int] = [self.normalize(image=SCREAMING_SNAKE_CASE_, mean=SCREAMING_SNAKE_CASE_, std=SCREAMING_SNAKE_CASE_ ) for image in images] UpperCAmelCase_: Dict = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) for image in images] UpperCAmelCase_: Any = {"""pixel_values""": images} return BatchFeature(data=SCREAMING_SNAKE_CASE_, tensor_type=SCREAMING_SNAKE_CASE_ )
147
1
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def UpperCamelCase_( snake_case__: bool = True , *snake_case__: str , **snake_case__: Optional[int] ) -> List[str]: if not is_tqdm_available(): raise ImportError('Accelerate\'s `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.' ) UpperCAmelCase__ = False if main_process_only: UpperCAmelCase__ = PartialState().local_process_index == 0 return _tqdm(*a__ , **a__ , disable=a__ )
371
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowercase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = BioGptTokenizer __SCREAMING_SNAKE_CASE = False def UpperCamelCase__ (self ) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] UpperCAmelCase__ = dict(zip(__a , range(len(__a ) ) ) ) UpperCAmelCase__ = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(__a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(__a ) ) def UpperCamelCase__ (self , __a ) -> Any: """simple docstring""" UpperCAmelCase__ = 'lower newer' UpperCAmelCase__ = 'lower newer' return input_text, output_text def UpperCamelCase__ (self ) -> Any: """simple docstring""" UpperCAmelCase__ = BioGptTokenizer(self.vocab_file , self.merges_file ) UpperCAmelCase__ = 'lower' UpperCAmelCase__ = ['low', 'er</w>'] UpperCAmelCase__ = tokenizer.tokenize(__a ) self.assertListEqual(__a , __a ) UpperCAmelCase__ = tokens + ['<unk>'] UpperCAmelCase__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__a ) , __a ) @slow def UpperCamelCase__ (self ) -> int: """simple docstring""" UpperCAmelCase__ = BioGptTokenizer.from_pretrained('microsoft/biogpt' ) UpperCAmelCase__ = tokenizer.encode('sequence builders' , add_special_tokens=__a ) UpperCAmelCase__ = tokenizer.encode('multi-sequence build' , add_special_tokens=__a ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(__a ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(__a , __a ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
335
0
'''simple docstring''' import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowercase_ = NewType("DataClass", Any) lowercase_ = NewType("DataClassType", Any) def lowerCAmelCase (__A): """simple docstring""" if isinstance(__A , __A): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''') def lowerCAmelCase (__A): """simple docstring""" _a = {str(__A): choice for choice in choices} return lambda __A: str_to_choice.get(__A , __A) def lowerCAmelCase (*, __A = None , __A = None , __A = dataclasses.MISSING , __A = dataclasses.MISSING , __A = None , **__A , ): """simple docstring""" if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _a = {} if aliases is not None: _a = aliases if help is not None: _a = help return dataclasses.field(metadata=__A , default=__A , default_factory=__A , **__A) class __A ( A ): '''simple docstring''' __lowerCamelCase : Iterable[DataClassType] def __init__(self , A , **A ) -> List[str]: """simple docstring""" if "formatter_class" not in kwargs: _a = ArgumentDefaultsHelpFormatter super().__init__(**A ) if dataclasses.is_dataclass(A ): _a = [dataclass_types] _a = list(A ) for dtype in self.dataclass_types: self._add_dataclass_arguments(A ) @staticmethod def a__ (A , A ) -> List[str]: """simple docstring""" _a = f'''--{field.name}''' _a = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , A ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) _a = kwargs.pop('''aliases''' , [] ) if isinstance(A , A ): _a = [aliases] _a = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(A , '''UnionType''' ) and isinstance(A , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(A ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f''' Problem encountered in field \'{field.name}\'.''' ) if type(A ) not in field.type.__args__: # filter `str` in Union _a = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _a = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _a = ( field.type.__args__[0] if isinstance(A , field.type.__args__[1] ) else field.type.__args__[1] ) _a = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _a = {} if origin_type is Literal or (isinstance(field.type , A ) and issubclass(field.type , A )): if origin_type is Literal: _a = field.type.__args__ else: _a = [x.value for x in field.type] _a = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: _a = field.default else: _a = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _a = copy(A ) # Hack because type=bool in argparse does not behave as we want. _a = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _a = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _a = default # This tells argparse we accept 0 or 1 value after --field_name _a = '''?''' # This is the value that will get picked if we do --field_name (without value) _a = True elif isclass(A ) and issubclass(A , A ): _a = field.type.__args__[0] _a = '''+''' if field.default_factory is not dataclasses.MISSING: _a = field.default_factory() elif field.default is dataclasses.MISSING: _a = True else: _a = field.type if field.default is not dataclasses.MISSING: _a = field.default elif field.default_factory is not dataclasses.MISSING: _a = field.default_factory() else: _a = True parser.add_argument(A , *A , **A ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _a = False parser.add_argument(f'''--no_{field.name}''' , action='''store_false''' , dest=field.name , **A ) def a__ (self , A ) -> int: """simple docstring""" if hasattr(A , '''_argument_group_name''' ): _a = self.add_argument_group(dtype._argument_group_name ) else: _a = self try: _a = get_type_hints(A ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(A ): _a = '''.'''.join(map(A , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(A ): if not field.init: continue _a = type_hints[field.name] self._parse_dataclass_field(A , A ) def a__ (self , A=None , A=False , A=True , A=None , A=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _a = [] if args_filename: args_files.append(Path(A ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _a = ArgumentParser() args_file_parser.add_argument(A , type=A , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) _a , _a = args_file_parser.parse_known_args(args=A ) _a = vars(A ).get(args_file_flag.lstrip('''-''' ) , A ) if cmd_args_file_paths: args_files.extend([Path(A ) for p in cmd_args_file_paths] ) _a = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _a = file_args + args if args is not None else file_args + sys.argv[1:] _a , _a = self.parse_known_args(args=A ) _a = [] for dtype in self.dataclass_types: _a = {f.name for f in dataclasses.fields(A ) if f.init} _a = {k: v for k, v in vars(A ).items() if k in keys} for k in keys: delattr(A , A ) _a = dtype(**A ) outputs.append(A ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(A ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def a__ (self , A , A = False ) -> Tuple[DataClass, ...]: """simple docstring""" _a = set(args.keys() ) _a = [] for dtype in self.dataclass_types: _a = {f.name for f in dataclasses.fields(A ) if f.init} _a = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _a = dtype(**A ) outputs.append(A ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(A )}''' ) return tuple(A ) def a__ (self , A , A = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(A ) , encoding='''utf-8''' ) as open_json_file: _a = json.loads(open_json_file.read() ) _a = self.parse_dict(A , allow_extra_keys=A ) return tuple(A ) def a__ (self , A , A = False ) -> Tuple[DataClass, ...]: """simple docstring""" _a = self.parse_dict(yaml.safe_load(Path(A ).read_text() ) , allow_extra_keys=A ) return tuple(A )
211
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowercase_ = logging.get_logger(__name__) lowercase_ = { "microsoft/resnet-50": "https://huggingface.co/microsoft/resnet-50/blob/main/config.json", } class __A ( A , A ): '''simple docstring''' __lowerCamelCase : Tuple = 'resnet' __lowerCamelCase : Any = ['basic', 'bottleneck'] def __init__(self , A=3 , A=64 , A=[256, 512, 1_024, 2_048] , A=[3, 4, 6, 3] , A="bottleneck" , A="relu" , A=False , A=None , A=None , **A , ) -> Dict: """simple docstring""" super().__init__(**A ) if layer_type not in self.layer_types: raise ValueError(f'''layer_type={layer_type} is not one of {','.join(self.layer_types )}''' ) _a = num_channels _a = embedding_size _a = hidden_sizes _a = depths _a = layer_type _a = hidden_act _a = downsample_in_first_stage _a = ['''stem'''] + [f'''stage{idx}''' for idx in range(1 , len(A ) + 1 )] _a , _a = get_aligned_output_features_output_indices( out_features=A , out_indices=A , stage_names=self.stage_names ) class __A ( A ): '''simple docstring''' __lowerCamelCase : Any = version.parse('1.11' ) @property def a__ (self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def a__ (self ) -> float: """simple docstring""" return 1E-3
211
1
"""simple docstring""" def _snake_case ( UpperCamelCase : int = 100 ): UpperCAmelCase : Optional[Any] = set() UpperCAmelCase : Union[str, Any] = 0 UpperCAmelCase : List[Any] = n + 1 # maximum limit for a in range(2 , UpperCamelCase ): for b in range(2 , UpperCamelCase ): UpperCAmelCase : Optional[Any] = a**b # calculates the current power collect_powers.add(UpperCamelCase ) # adds the result to the set return len(UpperCamelCase ) if __name__ == "__main__": print("Number of terms ", solution(int(str(input()).strip())))
76
"""simple docstring""" import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def _snake_case ( UpperCamelCase : Dict , UpperCamelCase : int , UpperCamelCase : List[str]=[] ): UpperCAmelCase : List[Any] = size[0] - overlap_pixels * 2 UpperCAmelCase : Dict = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels UpperCAmelCase : Union[str, Any] = np.ones((size_y, size_x) , dtype=np.uinta ) * 255 UpperCAmelCase : Any = np.pad(UpperCamelCase , mode="""linear_ramp""" , pad_width=UpperCamelCase , end_values=0 ) if "l" in remove_borders: UpperCAmelCase : Dict = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: UpperCAmelCase : Optional[Any] = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: UpperCAmelCase : Any = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: UpperCAmelCase : str = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def _snake_case ( UpperCamelCase : Optional[int] , UpperCamelCase : Dict , UpperCamelCase : Optional[Any] ): return max(UpperCamelCase , min(UpperCamelCase , UpperCamelCase ) ) def _snake_case ( UpperCamelCase : [int] , UpperCamelCase : [int] , UpperCamelCase : [int] ): return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def _snake_case ( UpperCamelCase : [int] , UpperCamelCase : int , UpperCamelCase : [int] ): UpperCAmelCase : Optional[Any] = list(UpperCamelCase ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap UpperCAmelCase : List[str] = clamp_rect(UpperCamelCase , [0, 0] , [image_size[0], image_size[1]] ) return rect def _snake_case ( UpperCamelCase : Optional[Any] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Any , UpperCamelCase : Dict ): UpperCAmelCase : Dict = Image.new("""RGB""" , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(UpperCamelCase , (original_slice, 0) ) return result def _snake_case ( UpperCamelCase : Tuple , UpperCamelCase : Union[str, Any] ): UpperCAmelCase : List[Any] = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) UpperCAmelCase : List[Any] = tile.crop(UpperCamelCase ) return tile def _snake_case ( UpperCamelCase : Optional[int] , UpperCamelCase : Optional[Any] ): UpperCAmelCase : Union[str, Any] = n % d return n - divisor class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 350 , ) -> List[Any]: '''simple docstring''' super().__init__( vae=_SCREAMING_SNAKE_CASE , text_encoder=_SCREAMING_SNAKE_CASE , tokenizer=_SCREAMING_SNAKE_CASE , unet=_SCREAMING_SNAKE_CASE , low_res_scheduler=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE , max_noise_level=_SCREAMING_SNAKE_CASE , ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : List[str] = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) UpperCAmelCase : Any = add_overlap_rect(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , image.size ) UpperCAmelCase : int = image.crop(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : int = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] UpperCAmelCase : Dict = translated_slice_x - (original_image_slice / 2) UpperCAmelCase : Optional[int] = max(0 , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = squeeze_tile(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[int] = to_input.size UpperCAmelCase : Dict = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) UpperCAmelCase : Optional[Any] = super(_SCREAMING_SNAKE_CASE , self ).__call__(image=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ).images[0] UpperCAmelCase : int = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) UpperCAmelCase : Union[str, Any] = unsqueeze_tile(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) UpperCAmelCase : int = [] if x == 0: remove_borders.append("""l""" ) elif crop_rect[2] == image.size[0]: remove_borders.append("""r""" ) if y == 0: remove_borders.append("""t""" ) elif crop_rect[3] == image.size[1]: remove_borders.append("""b""" ) UpperCAmelCase : Any = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=_SCREAMING_SNAKE_CASE ) , mode="""L""" , ) final_image.paste( _SCREAMING_SNAKE_CASE , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , _SCREAMING_SNAKE_CASE ) @torch.no_grad() def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 75 , _SCREAMING_SNAKE_CASE = 9.0 , _SCREAMING_SNAKE_CASE = 50 , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = 1 , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = 1 , _SCREAMING_SNAKE_CASE = 128 , _SCREAMING_SNAKE_CASE = 32 , _SCREAMING_SNAKE_CASE = 32 , ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : List[str] = Image.new("""RGB""" , (image.size[0] * 4, image.size[1] * 4) ) UpperCAmelCase : List[str] = math.ceil(image.size[0] / tile_size ) UpperCAmelCase : Optional[Any] = math.ceil(image.size[1] / tile_size ) UpperCAmelCase : Tuple = tcx * tcy UpperCAmelCase : Any = 0 for y in range(_SCREAMING_SNAKE_CASE ): for x in range(_SCREAMING_SNAKE_CASE ): self._process_tile( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , prompt=_SCREAMING_SNAKE_CASE , num_inference_steps=_SCREAMING_SNAKE_CASE , guidance_scale=_SCREAMING_SNAKE_CASE , noise_level=_SCREAMING_SNAKE_CASE , negative_prompt=_SCREAMING_SNAKE_CASE , num_images_per_prompt=_SCREAMING_SNAKE_CASE , eta=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , latents=_SCREAMING_SNAKE_CASE , ) current_count += 1 if callback is not None: callback({"""progress""": current_count / total_tile_count, """image""": final_image} ) return final_image def _snake_case ( ): # Run a demo UpperCAmelCase : Tuple = """stabilityai/stable-diffusion-x4-upscaler""" UpperCAmelCase : int = StableDiffusionTiledUpscalePipeline.from_pretrained(UpperCamelCase , revision="""fp16""" , torch_dtype=torch.floataa ) UpperCAmelCase : Any = pipe.to("""cuda""" ) UpperCAmelCase : Optional[int] = Image.open("""../../docs/source/imgs/diffusers_library.jpg""" ) def callback(UpperCamelCase : List[str] ): print(F"progress: {obj['progress']:.4f}" ) obj["image"].save("""diffusers_library_progress.jpg""" ) UpperCAmelCase : Any = pipe(image=UpperCamelCase , prompt="""Black font, white background, vector""" , noise_level=40 , callback=UpperCamelCase ) final_image.save("""diffusers_library.jpg""" ) if __name__ == "__main__": main()
76
1
'''simple docstring''' import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_ (snake_case__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase: Optional[Any] = LayoutLMTokenizer __UpperCamelCase: Dict = LayoutLMTokenizerFast __UpperCamelCase: List[Any] = True __UpperCamelCase: Union[str, Any] = True def _A ( self : str ): super().setUp() _UpperCAmelCase : Tuple = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] _UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def _A ( self : str , **A : List[str] ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **A ) def _A ( self : List[str] , A : List[Any] ): _UpperCAmelCase : int = "UNwant\u00E9d,running" _UpperCAmelCase : Optional[Any] = "unwanted, running" return input_text, output_text def _A ( self : int ): _UpperCAmelCase : List[Any] = self.tokenizer_class(self.vocab_file ) _UpperCAmelCase : List[Any] = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(A , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , [7, 4, 5, 10, 8, 9] ) def _A ( self : Union[str, Any] ): pass
31
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __snake_case : __lowerCamelCase = XGLMConfig __lowerCamelCase = {} __lowerCamelCase = """gelu""" def __init__( self , __UpperCamelCase , __UpperCamelCase=14 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=512 , __UpperCamelCase=0.0_2 , ) -> str: '''simple docstring''' snake_case__ : Any = parent snake_case__ : Optional[int] = batch_size snake_case__ : List[str] = seq_length snake_case__ : Optional[Any] = is_training snake_case__ : Optional[int] = use_input_mask snake_case__ : Any = use_labels snake_case__ : List[str] = vocab_size snake_case__ : List[Any] = d_model snake_case__ : List[str] = num_hidden_layers snake_case__ : Optional[int] = num_attention_heads snake_case__ : str = ffn_dim snake_case__ : Optional[Any] = activation_function snake_case__ : str = activation_dropout snake_case__ : int = attention_dropout snake_case__ : List[str] = max_position_embeddings snake_case__ : Optional[int] = initializer_range snake_case__ : List[str] = None snake_case__ : List[str] = 0 snake_case__ : Optional[int] = 2 snake_case__ : Union[str, Any] = 1 def __a ( self ) -> List[str]: '''simple docstring''' return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def __a ( self ) -> Optional[int]: '''simple docstring''' snake_case__ : List[str] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) snake_case__ : int = None if self.use_input_mask: snake_case__ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) snake_case__ : List[Any] = self.get_config() snake_case__ : Optional[int] = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def __a ( self ) -> Any: '''simple docstring''' return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=__UpperCamelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=__UpperCamelCase , ) def __a ( self ) -> Tuple: '''simple docstring''' snake_case__ : Any = self.prepare_config_and_inputs() ( ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ) : Tuple = config_and_inputs snake_case__ : Tuple = { 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class __snake_case ( _lowerCamelCase ,_lowerCamelCase ,unittest.TestCase ): __lowerCamelCase = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () __lowerCamelCase = (TFXGLMForCausalLM,) if is_tf_available() else () __lowerCamelCase = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __a ( self ) -> str: '''simple docstring''' snake_case__ : Union[str, Any] = TFXGLMModelTester(self ) snake_case__ : Optional[int] = ConfigTester(self , config_class=__UpperCamelCase , n_embd=37 ) def __a ( self ) -> int: '''simple docstring''' self.config_tester.run_common_tests() @slow def __a ( self ) -> Dict: '''simple docstring''' for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ : Any = TFXGLMModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def __a ( self ) -> Any: '''simple docstring''' super().test_resize_token_embeddings() @require_tf class __snake_case ( unittest.TestCase ): @slow def __a ( self , __UpperCamelCase=True ) -> int: '''simple docstring''' snake_case__ : Optional[Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) snake_case__ : Tuple = tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off snake_case__ : List[str] = [2, 268, 9865, 67, 11, 1988, 57252, 9865, 5, 984, 67, 1988, 213838, 1658, 53, 70446, 33, 6657, 278, 1581] # fmt: on snake_case__ : int = model.generate(__UpperCamelCase , do_sample=__UpperCamelCase , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , __UpperCamelCase ) @slow def __a ( self ) -> List[str]: '''simple docstring''' snake_case__ : Any = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) snake_case__ : Dict = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) snake_case__ : Any = tokenizer('Today is a nice day and' , return_tensors='tf' ) snake_case__ : Dict = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): snake_case__ : Optional[int] = model.generate(__UpperCamelCase , do_sample=__UpperCamelCase , seed=[7, 0] ) snake_case__ : int = tokenizer.decode(output_ids[0] , skip_special_tokens=__UpperCamelCase ) snake_case__ : str = ( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) @slow def __a ( self ) -> Dict: '''simple docstring''' snake_case__ : str = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) snake_case__ : Optional[int] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) snake_case__ : Any = 'left' # use different length sentences to test batching snake_case__ : int = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] snake_case__ : Any = tokenizer(__UpperCamelCase , return_tensors='tf' , padding=__UpperCamelCase ) snake_case__ : List[Any] = inputs['input_ids'] snake_case__ : List[str] = model.generate(input_ids=__UpperCamelCase , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) snake_case__ : Union[str, Any] = tokenizer(sentences[0] , return_tensors='tf' ).input_ids snake_case__ : str = model.generate(input_ids=__UpperCamelCase , max_new_tokens=12 ) snake_case__ : int = tokenizer(sentences[1] , return_tensors='tf' ).input_ids snake_case__ : Dict = model.generate(input_ids=__UpperCamelCase , max_new_tokens=12 ) snake_case__ : List[Any] = tokenizer.batch_decode(__UpperCamelCase , skip_special_tokens=__UpperCamelCase ) snake_case__ : Optional[int] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCamelCase ) snake_case__ : Dict = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCamelCase ) snake_case__ : Union[str, Any] = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) self.assertListEqual(__UpperCamelCase , [non_padded_sentence, padded_sentence] )
143
0
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __magic_name__( lowerCamelCase): return 1 / (1 + np.exp(-z)) def __magic_name__( lowerCamelCase, lowerCamelCase): return (-y * np.log(lowerCamelCase) - (1 - y) * np.log(1 - h)).mean() def __magic_name__( lowerCamelCase, lowerCamelCase, lowerCamelCase): __lowerCAmelCase = np.dot(lowerCamelCase, lowerCamelCase) return np.sum(y * scores - np.log(1 + np.exp(lowerCamelCase))) def __magic_name__( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=7_0_0_0_0): __lowerCAmelCase = np.zeros(x.shape[1]) for iterations in range(lowerCamelCase): __lowerCAmelCase = np.dot(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = sigmoid_function(lowerCamelCase) __lowerCAmelCase = np.dot(x.T, h - y) / y.size __lowerCAmelCase = theta - alpha * gradient # updating the weights __lowerCAmelCase = np.dot(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = sigmoid_function(lowerCamelCase) __lowerCAmelCase = cost_function(lowerCamelCase, lowerCamelCase) if iterations % 1_0_0 == 0: print(F"""loss: {j} \t""") # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": _UpperCAmelCase : Union[str, Any] = datasets.load_iris() _UpperCAmelCase : List[str] = iris.data[:, :2] _UpperCAmelCase : Dict = (iris.target != 0) * 1 _UpperCAmelCase : Tuple = 0.1 _UpperCAmelCase : str = logistic_reg(alpha, x, y, max_iterations=7_0_0_0_0) print("""theta: """, theta) # printing the theta i.e our weights vector def __magic_name__( lowerCamelCase): return sigmoid_function( np.dot(lowerCamelCase, lowerCamelCase)) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(1_0, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color="""b""", label="""0""") plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color="""r""", label="""1""") (_UpperCAmelCase) : Optional[Any] = (x[:, 0].min(), x[:, 0].max()) (_UpperCAmelCase) : str = (x[:, 1].min(), x[:, 1].max()) (_UpperCAmelCase) : List[Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) _UpperCAmelCase : Tuple = np.c_[xxa.ravel(), xxa.ravel()] _UpperCAmelCase : Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors="""black""") plt.legend() plt.show()
354
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class a__ ( __A ): """simple docstring""" __UpperCamelCase : Tuple = 'naver-clova-ix/donut-base-finetuned-docvqa' __UpperCamelCase : List[str] = ( 'This is a tool that answers a question about an document (pdf). It takes an input named `document` which ' 'should be the document containing the information, as well as a `question` that is the question about the ' 'document. It returns a text that contains the answer to the question.' ) __UpperCamelCase : Optional[int] = 'document_qa' __UpperCamelCase : Optional[int] = AutoProcessor __UpperCamelCase : Tuple = VisionEncoderDecoderModel __UpperCamelCase : Any = ['image', 'text'] __UpperCamelCase : Optional[Any] = ['text'] def __init__(self , *__lowercase , **__lowercase ): if not is_vision_available(): raise ValueError('''Pillow must be installed to use the DocumentQuestionAnsweringTool.''' ) super().__init__(*__lowercase , **__lowercase ) def _snake_case (self , __lowercase , __lowercase ): __lowerCAmelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' __lowerCAmelCase = task_prompt.replace('''{user_input}''' , __lowercase ) __lowerCAmelCase = self.pre_processor.tokenizer( __lowercase , add_special_tokens=__lowercase , return_tensors='''pt''' ).input_ids __lowerCAmelCase = self.pre_processor(__lowercase , return_tensors='''pt''' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _snake_case (self , __lowercase ): return self.model.generate( inputs['''pixel_values'''].to(self.device ) , decoder_input_ids=inputs['''decoder_input_ids'''].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__lowercase , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__lowercase , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__lowercase , ).sequences def _snake_case (self , __lowercase ): __lowerCAmelCase = self.pre_processor.batch_decode(__lowercase )[0] __lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '''''' ) __lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '''''' ) __lowerCAmelCase = re.sub(R'''<.*?>''' , '''''' , __lowercase , count=1 ).strip() # remove first task start token __lowerCAmelCase = self.pre_processor.tokenajson(__lowercase ) return sequence["answer"]
9
0
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy a_ = logging.get_logger(__name__) class _lowercase ( snake_case_ ): def __init__( self : List[Any] , snake_case : int , snake_case : int , snake_case : float , **snake_case : int ) -> int: """simple docstring""" UpperCamelCase_ : List[Any] = feature_size UpperCamelCase_ : Dict = sampling_rate UpperCamelCase_ : str = padding_value UpperCamelCase_ : List[Any] = kwargs.pop('padding_side' , 'right' ) UpperCamelCase_ : List[str] = kwargs.pop('return_attention_mask' , snake_case ) super().__init__(**snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , snake_case : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , snake_case : Union[bool, str, PaddingStrategy] = True , snake_case : Optional[int] = None , snake_case : bool = False , snake_case : Optional[int] = None , snake_case : Optional[bool] = None , snake_case : Optional[Union[str, TensorType]] = None , ) -> BatchFeature: """simple docstring""" if isinstance(snake_case , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCamelCase_ : Optional[int] = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' f" to this method that includes {self.model_input_names[0]}, but you provided" f" {list(processed_features.keys() )}" ) UpperCamelCase_ : str = processed_features[self.model_input_names[0]] UpperCamelCase_ : Union[str, Any] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(snake_case ) == 0: if return_attention_mask: UpperCamelCase_ : str = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCamelCase_ : Optional[int] = required_input[0] if isinstance(snake_case , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCamelCase_ : int = 0 while len(required_input[index] ) == 0: index += 1 if index < len(snake_case ): UpperCamelCase_ : List[str] = required_input[index][0] if return_tensors is None: if is_tf_tensor(snake_case ): UpperCamelCase_ : Dict = 'tf' elif is_torch_tensor(snake_case ): UpperCamelCase_ : List[Any] = 'pt' elif isinstance(snake_case , (int, float, list, tuple, np.ndarray) ): UpperCamelCase_ : List[Any] = 'np' else: raise ValueError( f"type of {first_element} unknown: {type(snake_case )}. " 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCamelCase_ : Tuple = to_numpy(snake_case ) else: UpperCamelCase_ : Dict = [to_numpy(snake_case ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCamelCase_ : Optional[Any] = self._get_padding_strategies(padding=snake_case , max_length=snake_case ) UpperCamelCase_ : Dict = processed_features[self.model_input_names[0]] UpperCamelCase_ : Any = len(snake_case ) if not all(len(snake_case ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) UpperCamelCase_ : Any = [] for i in range(snake_case ): UpperCamelCase_ : Union[str, Any] = {k: v[i] for k, v in processed_features.items()} # truncation UpperCamelCase_ : Any = self._truncate( snake_case , max_length=snake_case , pad_to_multiple_of=snake_case , truncation=snake_case , ) truncated_inputs.append(snake_case ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCamelCase_ : List[Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCamelCase_ : int = PaddingStrategy.MAX_LENGTH UpperCamelCase_ : Union[str, Any] = {} for i in range(snake_case ): # padding UpperCamelCase_ : List[str] = self._pad( truncated_inputs[i] , max_length=snake_case , padding_strategy=snake_case , pad_to_multiple_of=snake_case , return_attention_mask=snake_case , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCamelCase_ : int = [] if value.dtype is np.dtype(np.floataa ): UpperCamelCase_ : Union[str, Any] = value.astype(np.floataa ) batch_outputs[key].append(snake_case ) return BatchFeature(snake_case , tensor_type=snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Union[Dict[str, np.ndarray], BatchFeature] , snake_case : Optional[int] = None , snake_case : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , snake_case : Optional[int] = None , snake_case : Optional[bool] = None , ) -> dict: """simple docstring""" UpperCamelCase_ : List[Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCamelCase_ : Optional[int] = len(snake_case ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCamelCase_ : List[str] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCamelCase_ : Tuple = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(snake_case ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCamelCase_ : str = np.ones(len(snake_case ) , dtype=np.intaa ) if needs_to_be_padded: UpperCamelCase_ : str = max_length - len(snake_case ) if self.padding_side == "right": if return_attention_mask: UpperCamelCase_ : Tuple = np.pad( processed_features['attention_mask'] , (0, difference) ) UpperCamelCase_ : str = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCamelCase_ : List[Any] = np.pad( snake_case , snake_case , 'constant' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCamelCase_ : Tuple = np.pad( processed_features['attention_mask'] , (difference, 0) ) UpperCamelCase_ : List[Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCamelCase_ : Dict = np.pad( snake_case , snake_case , 'constant' , constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Union[Dict[str, np.ndarray], BatchFeature] , snake_case : Optional[int] = None , snake_case : Optional[int] = None , snake_case : Optional[bool] = None , ) -> Any: """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) UpperCamelCase_ : str = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCamelCase_ : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCamelCase_ : Optional[Any] = len(snake_case ) > max_length if needs_to_be_truncated: UpperCamelCase_ : str = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCamelCase_ : Optional[int] = processed_features['attention_mask'][:max_length] return processed_features def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : str=False , snake_case : Dict=None ) -> int: """simple docstring""" if padding is not False: if padding is True: UpperCamelCase_ : Any = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(snake_case , snake_case ): UpperCamelCase_ : List[Any] = PaddingStrategy(snake_case ) elif isinstance(snake_case , snake_case ): UpperCamelCase_ : Optional[int] = padding else: UpperCamelCase_ : Optional[Any] = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( f"When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
175
def __lowercase ( lowerCamelCase : str ): UpperCamelCase_ : Dict = 0 for ch in input_str: UpperCamelCase_ : Tuple = ord(lowerCamelCase ) UpperCamelCase_ : str = pow(2 , lowerCamelCase ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
175
1
"""simple docstring""" import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging _lowerCAmelCase : Tuple = logging.get_logger(__name__) def __snake_case ( SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = r"""\w+[.]\d+""" _UpperCAmelCase : Union[str, Any] = re.findall(_lowerCAmelCase , _lowerCAmelCase ) for pat in pats: _UpperCAmelCase : Any = key.replace(_lowerCAmelCase , "_".join(pat.split("." ) ) ) return key def __snake_case ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple ) -> Any: '''simple docstring''' _UpperCAmelCase : List[Any] = pt_tuple_key[:-1] + ("""scale""",) if ( any("norm" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): _UpperCAmelCase : Tuple = pt_tuple_key[:-1] + ("""scale""",) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: _UpperCAmelCase : Dict = pt_tuple_key[:-1] + ("""scale""",) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: _UpperCAmelCase : str = pt_tuple_key[:-1] + ("""embedding""",) return renamed_pt_tuple_key, pt_tensor # conv layer _UpperCAmelCase : Optional[int] = pt_tuple_key[:-1] + ("""kernel""",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: _UpperCAmelCase : Optional[int] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer _UpperCAmelCase : List[Any] = pt_tuple_key[:-1] + ("""kernel""",) if pt_tuple_key[-1] == "weight": _UpperCAmelCase : Optional[int] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight _UpperCAmelCase : int = pt_tuple_key[:-1] + ("""weight""",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias _UpperCAmelCase : int = pt_tuple_key[:-1] + ("""bias""",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def __snake_case ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any=42 ) -> List[Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params _UpperCAmelCase : Union[str, Any] = flax_model.init_weights(PRNGKey(_lowerCAmelCase ) ) _UpperCAmelCase : Optional[Any] = flatten_dict(_lowerCAmelCase ) _UpperCAmelCase : Optional[int] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): _UpperCAmelCase : Any = rename_key(_lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = tuple(renamed_pt_key.split("." ) ) # Correctly rename weight parameters _UpperCAmelCase : int = rename_key_and_reshape_tensor(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' f'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # also add unexpected weight so that warning is thrown _UpperCAmelCase : int = jnp.asarray(_lowerCAmelCase ) return unflatten_dict(_lowerCAmelCase )
371
"""simple docstring""" def __snake_case ( SCREAMING_SNAKE_CASE__ : List[str] ) -> str: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] while len(SCREAMING_SNAKE_CASE__ ) > 1: _UpperCAmelCase , _UpperCAmelCase : int = min(SCREAMING_SNAKE_CASE__ ), max(SCREAMING_SNAKE_CASE__ ) start.append(SCREAMING_SNAKE_CASE__ ) end.append(SCREAMING_SNAKE_CASE__ ) collection.remove(SCREAMING_SNAKE_CASE__ ) collection.remove(SCREAMING_SNAKE_CASE__ ) end.reverse() return start + collection + end if __name__ == "__main__": _lowerCAmelCase : int = input("Enter numbers separated by a comma:\n").strip() _lowerCAmelCase : List[str] = [int(item) for item in user_input.split(",")] print(*merge_sort(unsorted), sep=",")
202
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: lowerCAmelCase = None lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase = { """vocab_file""": { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/spiece.model""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/spiece.model""", }, """tokenizer_file""": { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json""", }, } lowerCAmelCase = { """google/fnet-base""": 5_12, """google/fnet-large""": 5_12, } lowerCAmelCase = """▁""" class A_ ( A__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = ["""input_ids""", """token_type_ids"""] SCREAMING_SNAKE_CASE_ = FNetTokenizer def __init__( self :List[Any] , lowerCamelCase_ :int=None , lowerCamelCase_ :Dict=None , lowerCamelCase_ :int=False , lowerCamelCase_ :str=True , lowerCamelCase_ :Optional[Any]=True , lowerCamelCase_ :List[str]="<unk>" , lowerCamelCase_ :Optional[int]="[SEP]" , lowerCamelCase_ :List[str]="<pad>" , lowerCamelCase_ :List[str]="[CLS]" , lowerCamelCase_ :Tuple="[MASK]" , **lowerCamelCase_ :Optional[Any] , ): """simple docstring""" lowerCamelCase__ : Dict =( AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ , normalized=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else mask_token ) super().__init__( lowerCamelCase_ , tokenizer_file=lowerCamelCase_ , do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , **lowerCamelCase_ , ) lowerCamelCase__ : Optional[int] =do_lower_case lowerCamelCase__ : Dict =remove_space lowerCamelCase__ : Optional[int] =keep_accents lowerCamelCase__ : str =vocab_file lowerCamelCase__ : Tuple =False if not self.vocab_file else True def UpperCAmelCase__ ( self :Union[str, Any] , lowerCamelCase_ :List[int] , lowerCamelCase_ :Optional[List[int]] = None ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =[self.sep_token_id] lowerCamelCase__ : List[str] =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase__ ( self :str , lowerCamelCase_ :List[int] , lowerCamelCase_ :Optional[List[int]] = None ): """simple docstring""" lowerCamelCase__ : List[Any] =[self.sep_token_id] lowerCamelCase__ : Tuple =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self :Union[str, Any] , lowerCamelCase_ :str , lowerCamelCase_ :Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : List[str] =os.path.join( lowerCamelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase_ ): copyfile(self.vocab_file , lowerCamelCase_ ) return (out_vocab_file,)
126
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A_ ( A__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ = AudioLDMPipeline SCREAMING_SNAKE_CASE_ = TEXT_TO_AUDIO_PARAMS SCREAMING_SNAKE_CASE_ = TEXT_TO_AUDIO_BATCH_PARAMS SCREAMING_SNAKE_CASE_ = frozenset( [ """num_inference_steps""", """num_waveforms_per_prompt""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCAmelCase__ ( self :Optional[int] ): """simple docstring""" torch.manual_seed(0 ) lowerCamelCase__ : Optional[int] =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=(32, 64) , class_embed_type='simple_projection' , projection_class_embeddings_input_dim=32 , class_embeddings_concat=lowerCamelCase_ , ) lowerCamelCase__ : Any =DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase_ , set_alpha_to_one=lowerCamelCase_ , ) torch.manual_seed(0 ) lowerCamelCase__ : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCamelCase__ : Union[str, Any] =ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , projection_dim=32 , ) lowerCamelCase__ : Any =ClapTextModelWithProjection(lowerCamelCase_ ) lowerCamelCase__ : str =RobertaTokenizer.from_pretrained('hf-internal-testing/tiny-random-roberta' , model_max_length=77 ) lowerCamelCase__ : int =SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=16_000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=lowerCamelCase_ , ) lowerCamelCase__ : Optional[Any] =SpeechTaHifiGan(lowerCamelCase_ ) lowerCamelCase__ : Optional[int] ={ 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'vocoder': vocoder, } return components def UpperCAmelCase__ ( self :Dict , lowerCamelCase_ :Dict , lowerCamelCase_ :List[Any]=0 ): """simple docstring""" if str(lowerCamelCase_ ).startswith('mps' ): lowerCamelCase__ : Optional[int] =torch.manual_seed(lowerCamelCase_ ) else: lowerCamelCase__ : int =torch.Generator(device=lowerCamelCase_ ).manual_seed(lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] ={ 'prompt': 'A hammer hitting a wooden surface', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, } return inputs def UpperCAmelCase__ ( self :Tuple ): """simple docstring""" lowerCamelCase__ : Optional[Any] ='cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase__ : Any =self.get_dummy_components() lowerCamelCase__ : List[str] =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : int =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =audioldm_pipe(**lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) == 256 lowerCamelCase__ : int =audio[:10] lowerCamelCase__ : Dict =np.array( [-0.00_50, 0.00_50, -0.00_60, 0.00_33, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_33] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def UpperCAmelCase__ ( self :str ): """simple docstring""" lowerCamelCase__ : List[str] =self.get_dummy_components() lowerCamelCase__ : str =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Optional[int] =audioldm_pipe.to(lowerCamelCase_ ) lowerCamelCase__ : Tuple =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : List[str] =3 * [inputs['prompt']] # forward lowerCamelCase__ : int =audioldm_pipe(**lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =output.audios[0] lowerCamelCase__ : List[Any] =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : List[Any] =3 * [inputs.pop('prompt' )] lowerCamelCase__ : List[str] =audioldm_pipe.tokenizer( lowerCamelCase_ , padding='max_length' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase_ , return_tensors='pt' , ) lowerCamelCase__ : Optional[int] =text_inputs['input_ids'].to(lowerCamelCase_ ) lowerCamelCase__ : str =audioldm_pipe.text_encoder( lowerCamelCase_ , ) lowerCamelCase__ : Any =prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state lowerCamelCase__ : Tuple =F.normalize(lowerCamelCase_ , dim=-1 ) lowerCamelCase__ : int =prompt_embeds # forward lowerCamelCase__ : List[str] =audioldm_pipe(**lowerCamelCase_ ) lowerCamelCase__ : Tuple =output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def UpperCAmelCase__ ( self :List[Any] ): """simple docstring""" lowerCamelCase__ : List[str] =self.get_dummy_components() lowerCamelCase__ : Union[str, Any] =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =audioldm_pipe.to(lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Optional[int] =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : Any =3 * ['this is a negative prompt'] lowerCamelCase__ : Union[str, Any] =negative_prompt lowerCamelCase__ : Union[str, Any] =3 * [inputs['prompt']] # forward lowerCamelCase__ : List[Any] =audioldm_pipe(**lowerCamelCase_ ) lowerCamelCase__ : Dict =output.audios[0] lowerCamelCase__ : Optional[int] =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : List[Any] =3 * [inputs.pop('prompt' )] lowerCamelCase__ : Union[str, Any] =[] for p in [prompt, negative_prompt]: lowerCamelCase__ : List[str] =audioldm_pipe.tokenizer( lowerCamelCase_ , padding='max_length' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase_ , return_tensors='pt' , ) lowerCamelCase__ : Optional[Any] =text_inputs['input_ids'].to(lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =audioldm_pipe.text_encoder( lowerCamelCase_ , ) lowerCamelCase__ : List[str] =text_embeds.text_embeds # additional L_2 normalization over each hidden-state lowerCamelCase__ : List[Any] =F.normalize(lowerCamelCase_ , dim=-1 ) embeds.append(lowerCamelCase_ ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] =embeds # forward lowerCamelCase__ : Any =audioldm_pipe(**lowerCamelCase_ ) lowerCamelCase__ : Dict =output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def UpperCAmelCase__ ( self :int ): """simple docstring""" lowerCamelCase__ : List[Any] ='cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase__ : List[Any] =self.get_dummy_components() lowerCamelCase__ : List[Any] =PNDMScheduler(skip_prk_steps=lowerCamelCase_ ) lowerCamelCase__ : int =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : str =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] ='egg cracking' lowerCamelCase__ : int =audioldm_pipe(**lowerCamelCase_ , negative_prompt=lowerCamelCase_ ) lowerCamelCase__ : Dict =output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) == 256 lowerCamelCase__ : Optional[Any] =audio[:10] lowerCamelCase__ : str =np.array( [-0.00_51, 0.00_50, -0.00_60, 0.00_34, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_32] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def UpperCAmelCase__ ( self :str ): """simple docstring""" lowerCamelCase__ : Tuple ='cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase__ : int =self.get_dummy_components() lowerCamelCase__ : Optional[Any] =PNDMScheduler(skip_prk_steps=lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Union[str, Any] ='A hammer hitting a wooden surface' # test num_waveforms_per_prompt=1 (default) lowerCamelCase__ : int =audioldm_pipe(lowerCamelCase_ , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts lowerCamelCase__ : Union[str, Any] =2 lowerCamelCase__ : Optional[int] =audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt lowerCamelCase__ : Optional[Any] =2 lowerCamelCase__ : List[Any] =audioldm_pipe(lowerCamelCase_ , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase_ ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts lowerCamelCase__ : List[Any] =2 lowerCamelCase__ : str =audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase_ ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : str ='cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase__ : str =self.get_dummy_components() lowerCamelCase__ : Tuple =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Optional[int] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Dict =audioldm_pipe.vocoder.config.sampling_rate lowerCamelCase__ : int =self.get_dummy_inputs(lowerCamelCase_ ) lowerCamelCase__ : Any =audioldm_pipe(audio_length_in_s=0.0_16 , **lowerCamelCase_ ) lowerCamelCase__ : List[Any] =output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) / vocoder_sampling_rate == 0.0_16 lowerCamelCase__ : Optional[Any] =audioldm_pipe(audio_length_in_s=0.0_32 , **lowerCamelCase_ ) lowerCamelCase__ : Any =output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) / vocoder_sampling_rate == 0.0_32 def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : Any =self.get_dummy_components() lowerCamelCase__ : List[Any] =AudioLDMPipeline(**lowerCamelCase_ ) lowerCamelCase__ : Tuple =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Any =['hey'] lowerCamelCase__ : Dict =audioldm_pipe(lowerCamelCase_ , num_inference_steps=1 ) lowerCamelCase__ : int =output.audios.shape assert audio_shape == (1, 256) lowerCamelCase__ : Union[str, Any] =audioldm_pipe.vocoder.config config.model_in_dim *= 2 lowerCamelCase__ : Tuple =SpeechTaHifiGan(lowerCamelCase_ ).to(lowerCamelCase_ ) lowerCamelCase__ : Tuple =audioldm_pipe(lowerCamelCase_ , num_inference_steps=1 ) lowerCamelCase__ : List[str] =output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def UpperCAmelCase__ ( self :str ): """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCamelCase_ ) def UpperCAmelCase__ ( self :Any ): """simple docstring""" self._test_inference_batch_single_identical(test_mean_pixel_difference=lowerCamelCase_ ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def UpperCAmelCase__ ( self :Optional[int] ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase_ ) @slow class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :str ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self :Optional[int] , lowerCamelCase_ :Any , lowerCamelCase_ :List[str]="cpu" , lowerCamelCase_ :Tuple=torch.floataa , lowerCamelCase_ :str=0 ): """simple docstring""" lowerCamelCase__ : Tuple =torch.Generator(device=lowerCamelCase_ ).manual_seed(lowerCamelCase_ ) lowerCamelCase__ : List[str] =np.random.RandomState(lowerCamelCase_ ).standard_normal((1, 8, 128, 16) ) lowerCamelCase__ : Tuple =torch.from_numpy(lowerCamelCase_ ).to(device=lowerCamelCase_ , dtype=lowerCamelCase_ ) lowerCamelCase__ : Dict ={ 'prompt': 'A hammer hitting a wooden surface', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 2.5, } return inputs def UpperCAmelCase__ ( self :List[Any] ): """simple docstring""" lowerCamelCase__ : int =AudioLDMPipeline.from_pretrained('cvssp/audioldm' ) lowerCamelCase__ : List[Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Dict =self.get_inputs(lowerCamelCase_ ) lowerCamelCase__ : Any =25 lowerCamelCase__ : List[Any] =audioldm_pipe(**lowerCamelCase_ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) == 81_920 lowerCamelCase__ : Dict =audio[77_230:77_240] lowerCamelCase__ : List[Any] =np.array( [-0.48_84, -0.46_07, 0.00_23, 0.50_07, 0.58_96, 0.51_51, 0.38_13, -0.02_08, -0.36_87, -0.43_15] ) lowerCamelCase__ : Any =np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def UpperCAmelCase__ ( self :Union[str, Any] ): """simple docstring""" lowerCamelCase__ : Any =AudioLDMPipeline.from_pretrained('cvssp/audioldm' ) lowerCamelCase__ : int =LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) lowerCamelCase__ : List[Any] =audioldm_pipe.to(lowerCamelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =self.get_inputs(lowerCamelCase_ ) lowerCamelCase__ : List[Any] =audioldm_pipe(**lowerCamelCase_ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase_ ) == 81_920 lowerCamelCase__ : Union[str, Any] =audio[27_780:27_790] lowerCamelCase__ : Dict =np.array([-0.21_31, -0.08_73, -0.01_24, -0.01_89, 0.05_69, 0.13_73, 0.18_83, 0.28_86, 0.32_97, 0.22_12] ) lowerCamelCase__ : Tuple =np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
126
1
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" monkeypatch.setattr('''datasets.utils.deprecation_utils._emitted_deprecation_warnings''' ,set() ) @pytest.fixture def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" class A__ : """simple docstring""" def __init__( self , __snake_case ): snake_case = metric_id class A__ : """simple docstring""" __magic_name__ = [MetricMock(snake_case__ ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def a_ ( self ): return self._metrics monkeypatch.setattr('''datasets.inspect.huggingface_hub''' ,HfhMock() ) @pytest.mark.parametrize( '''func, args''' ,[(load_metric, ('''metrics/mse''',)), (list_metrics, ()), (inspect_metric, ('''metrics/mse''', '''tmp_path'''))] ) def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" if "tmp_path" in args: snake_case = tuple(arg if arg != '''tmp_path''' else tmp_path for arg in args ) with pytest.warns(a__ ,match='''https://huggingface.co/docs/evaluate''' ): func(*a__ )
361
from __future__ import annotations import time _SCREAMING_SNAKE_CASE : List[Any] = list[tuple[int, int]] _SCREAMING_SNAKE_CASE : Any = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _SCREAMING_SNAKE_CASE : Any = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class A__ : """simple docstring""" def __init__( self , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ): snake_case = pos_x snake_case = pos_y snake_case = (pos_y, pos_x) snake_case = goal_x snake_case = goal_y snake_case = parent class A__ : """simple docstring""" def __init__( self , __snake_case , __snake_case ): snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , __snake_case ) snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , __snake_case ) snake_case = [self.start] snake_case = False def a_ ( self ): while self.node_queue: snake_case = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: snake_case = True return self.retrace_path(__snake_case ) snake_case = self.get_successors(__snake_case ) for node in successors: self.node_queue.append(__snake_case ) if not self.reached: return [self.start.pos] return None def a_ ( self , __snake_case ): snake_case = [] for action in delta: snake_case = parent.pos_x + action[1] snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(__snake_case ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(__snake_case , __snake_case , self.target.pos_y , self.target.pos_x , __snake_case ) ) return successors def a_ ( self , __snake_case ): snake_case = node snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) snake_case = current_node.parent path.reverse() return path class A__ : """simple docstring""" def __init__( self , __snake_case , __snake_case ): snake_case = BreadthFirstSearch(__snake_case , __snake_case ) snake_case = BreadthFirstSearch(__snake_case , __snake_case ) snake_case = False def a_ ( self ): while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: snake_case = self.fwd_bfs.node_queue.pop(0 ) snake_case = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: snake_case = True return self.retrace_bidirectional_path( __snake_case , __snake_case ) snake_case = current_bwd_node snake_case = current_fwd_node snake_case = { self.fwd_bfs: self.fwd_bfs.get_successors(__snake_case ), self.bwd_bfs: self.bwd_bfs.get_successors(__snake_case ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(__snake_case ) if not self.reached: return [self.fwd_bfs.start.pos] return None def a_ ( self , __snake_case , __snake_case ): snake_case = self.fwd_bfs.retrace_path(__snake_case ) snake_case = self.bwd_bfs.retrace_path(__snake_case ) bwd_path.pop() bwd_path.reverse() snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() _SCREAMING_SNAKE_CASE : Optional[Any] = (0, 0) _SCREAMING_SNAKE_CASE : List[Any] = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _SCREAMING_SNAKE_CASE : List[Any] = time.time() _SCREAMING_SNAKE_CASE : List[Any] = BreadthFirstSearch(init, goal) _SCREAMING_SNAKE_CASE : List[str] = bfs.search() _SCREAMING_SNAKE_CASE : int = time.time() - start_bfs_time print("Unidirectional BFS computation time : ", bfs_time) _SCREAMING_SNAKE_CASE : Any = time.time() _SCREAMING_SNAKE_CASE : Union[str, Any] = BidirectionalBreadthFirstSearch(init, goal) _SCREAMING_SNAKE_CASE : Union[str, Any] = bd_bfs.search() _SCREAMING_SNAKE_CASE : Tuple = time.time() - start_bd_bfs_time print("Bidirectional BFS computation time : ", bd_bfs_time)
213
0
from __future__ import annotations def UpperCamelCase( __UpperCamelCase : list ): if len(__UpperCamelCase ) == 0: return [] lowerCAmelCase_ , lowerCAmelCase_ : List[str] = min(__UpperCamelCase ), max(__UpperCamelCase ) lowerCAmelCase_ : List[Any] = int(max_value - min_value ) + 1 lowerCAmelCase_ : list[list] = [[] for _ in range(__UpperCamelCase )] for i in my_list: buckets[int(i - min_value )].append(__UpperCamelCase ) return [v for bucket in buckets for v in sorted(__UpperCamelCase )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
103
"""simple docstring""" import argparse import torch from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert from transformers.utils import logging logging.set_verbosity_info() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" lowerCAmelCase__ :str = BertConfig.from_json_file(_SCREAMING_SNAKE_CASE ) print(F"Building PyTorch model from configuration: {config}" ) lowerCAmelCase__ :int = BertForPreTraining(_SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_bert(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--bert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained BERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __A = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
293
0
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig _lowerCamelCase : int = logging.get_logger(__name__) # General docstring _lowerCamelCase : List[Any] = """PoolFormerConfig""" # Base docstring _lowerCamelCase : Optional[int] = """sail/poolformer_s12""" _lowerCamelCase : Union[str, Any] = [1, 5_12, 7, 7] # Image classification docstring _lowerCamelCase : Tuple = """sail/poolformer_s12""" _lowerCamelCase : Tuple = """tabby, tabby cat""" _lowerCamelCase : str = [ """sail/poolformer_s12""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def __lowerCamelCase (UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict = 0.0 , UpperCAmelCase__ : Tuple = False ): if drop_prob == 0.0 or not training: return input SCREAMING_SNAKE_CASE = 1 - drop_prob SCREAMING_SNAKE_CASE = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets SCREAMING_SNAKE_CASE = keep_prob + torch.rand(UpperCAmelCase__ , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize SCREAMING_SNAKE_CASE = input.div(UpperCAmelCase__ ) * random_tensor return output class lowercase ( nn.Module ): def __init__( self : Tuple , _UpperCamelCase : Union[str, Any] = None ) -> Optional[Any]: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = drop_prob def __snake_case( self : Union[str, Any] , _UpperCamelCase : List[Any] ) -> Union[str, Any]: '''simple docstring''' return drop_path(__lowercase , self.drop_prob , self.training ) def __snake_case( self : List[str] ) -> Optional[Any]: '''simple docstring''' return "p={}".format(self.drop_prob ) class lowercase ( nn.Module ): def __init__( self : List[str] , _UpperCamelCase : str , _UpperCamelCase : Dict , _UpperCamelCase : List[Any] , _UpperCamelCase : Dict , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[str]=None ) -> Any: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = patch_size if isinstance(__lowercase , collections.abc.Iterable ) else (patch_size, patch_size) SCREAMING_SNAKE_CASE = stride if isinstance(__lowercase , collections.abc.Iterable ) else (stride, stride) SCREAMING_SNAKE_CASE = padding if isinstance(__lowercase , collections.abc.Iterable ) else (padding, padding) SCREAMING_SNAKE_CASE = nn.Convad(__lowercase , __lowercase , kernel_size=__lowercase , stride=__lowercase , padding=__lowercase ) SCREAMING_SNAKE_CASE = norm_layer(__lowercase ) if norm_layer else nn.Identity() def __snake_case( self : Optional[int] , _UpperCamelCase : List[str] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = self.projection(__lowercase ) SCREAMING_SNAKE_CASE = self.norm(__lowercase ) return embeddings class lowercase ( nn.GroupNorm ): def __init__( self : Optional[Any] , _UpperCamelCase : int , **_UpperCamelCase : Union[str, Any] ) -> int: '''simple docstring''' super().__init__(1 , __lowercase , **__lowercase ) class lowercase ( nn.Module ): def __init__( self : Any , _UpperCamelCase : Any ) -> str: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = nn.AvgPoolad(__lowercase , stride=1 , padding=pool_size // 2 , count_include_pad=__lowercase ) def __snake_case( self : List[Any] , _UpperCamelCase : str ) -> Union[str, Any]: '''simple docstring''' return self.pool(__lowercase ) - hidden_states class lowercase ( nn.Module ): def __init__( self : int , _UpperCamelCase : Dict , _UpperCamelCase : Optional[int] , _UpperCamelCase : Dict , _UpperCamelCase : Dict ) -> Optional[Any]: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = nn.Convad(__lowercase , __lowercase , 1 ) SCREAMING_SNAKE_CASE = nn.Convad(__lowercase , __lowercase , 1 ) SCREAMING_SNAKE_CASE = PoolFormerDropPath(__lowercase ) if isinstance(config.hidden_act , __lowercase ): SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: SCREAMING_SNAKE_CASE = config.hidden_act def __snake_case( self : Optional[int] , _UpperCamelCase : List[Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.conva(__lowercase ) SCREAMING_SNAKE_CASE = self.act_fn(__lowercase ) SCREAMING_SNAKE_CASE = self.drop(__lowercase ) SCREAMING_SNAKE_CASE = self.conva(__lowercase ) SCREAMING_SNAKE_CASE = self.drop(__lowercase ) return hidden_states class lowercase ( nn.Module ): def __init__( self : Optional[int] , _UpperCamelCase : List[str] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Any , _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] , _UpperCamelCase : Optional[int] ) -> int: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = PoolFormerPooling(__lowercase ) SCREAMING_SNAKE_CASE = PoolFormerOutput(__lowercase , __lowercase , __lowercase , __lowercase ) SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(__lowercase ) SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(__lowercase ) # Useful for training neural nets SCREAMING_SNAKE_CASE = PoolFormerDropPath(__lowercase ) if drop_path > 0.0 else nn.Identity() SCREAMING_SNAKE_CASE = config.use_layer_scale if config.use_layer_scale: SCREAMING_SNAKE_CASE = nn.Parameter( config.layer_scale_init_value * torch.ones((__lowercase) ) , requires_grad=__lowercase ) SCREAMING_SNAKE_CASE = nn.Parameter( config.layer_scale_init_value * torch.ones((__lowercase) ) , requires_grad=__lowercase ) def __snake_case( self : List[Any] , _UpperCamelCase : Tuple ) -> Tuple: '''simple docstring''' if self.use_layer_scale: SCREAMING_SNAKE_CASE = self.pooling(self.before_norm(__lowercase ) ) SCREAMING_SNAKE_CASE = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection SCREAMING_SNAKE_CASE = hidden_states + self.drop_path(__lowercase ) SCREAMING_SNAKE_CASE = () SCREAMING_SNAKE_CASE = self.output(self.after_norm(__lowercase ) ) SCREAMING_SNAKE_CASE = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection SCREAMING_SNAKE_CASE = hidden_states + self.drop_path(__lowercase ) SCREAMING_SNAKE_CASE = (output,) + outputs return outputs else: SCREAMING_SNAKE_CASE = self.drop_path(self.pooling(self.before_norm(__lowercase ) ) ) # First residual connection SCREAMING_SNAKE_CASE = pooling_output + hidden_states SCREAMING_SNAKE_CASE = () # Second residual connection inside the PoolFormerOutput block SCREAMING_SNAKE_CASE = self.drop_path(self.output(self.after_norm(__lowercase ) ) ) SCREAMING_SNAKE_CASE = hidden_states + layer_output SCREAMING_SNAKE_CASE = (output,) + outputs return outputs class lowercase ( nn.Module ): def __init__( self : Any , _UpperCamelCase : Any ) -> List[str]: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = config # stochastic depth decay rule SCREAMING_SNAKE_CASE = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings SCREAMING_SNAKE_CASE = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) SCREAMING_SNAKE_CASE = nn.ModuleList(__lowercase ) # Transformer blocks SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers SCREAMING_SNAKE_CASE = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( __lowercase , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(__lowercase ) ) SCREAMING_SNAKE_CASE = nn.ModuleList(__lowercase ) def __snake_case( self : str , _UpperCamelCase : str , _UpperCamelCase : Dict=False , _UpperCamelCase : str=True ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = () if output_hidden_states else None SCREAMING_SNAKE_CASE = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = layers # Get patch embeddings from hidden_states SCREAMING_SNAKE_CASE = embedding_layer(__lowercase ) # Send the embeddings through the blocks for _, blk in enumerate(__lowercase ): SCREAMING_SNAKE_CASE = blk(__lowercase ) SCREAMING_SNAKE_CASE = layer_outputs[0] if output_hidden_states: SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=__lowercase , hidden_states=__lowercase ) class lowercase ( __A ): lowercase__ : int = PoolFormerConfig lowercase__ : Union[str, Any] = 'poolformer' lowercase__ : Tuple = 'pixel_values' lowercase__ : Tuple = True def __snake_case( self : Any , _UpperCamelCase : Any ) -> str: '''simple docstring''' if isinstance(__lowercase , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(__lowercase , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def __snake_case( self : List[str] , _UpperCamelCase : List[str] , _UpperCamelCase : Optional[int]=False ) -> Dict: '''simple docstring''' if isinstance(__lowercase , __lowercase ): SCREAMING_SNAKE_CASE = value _lowerCamelCase : Any = r""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ _lowerCamelCase : List[Any] = r""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`PoolFormerImageProcessor.__call__`] for details. """ @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , __A , ) class lowercase ( __A ): def __init__( self : List[Any] , _UpperCamelCase : Dict ) -> int: '''simple docstring''' super().__init__(__lowercase ) SCREAMING_SNAKE_CASE = config SCREAMING_SNAKE_CASE = PoolFormerEncoder(__lowercase ) # Initialize weights and apply final processing self.post_init() def __snake_case( self : List[str] ) -> Any: '''simple docstring''' return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(__lowercase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__lowercase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __snake_case( self : Optional[Any] , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : List[str] = None , _UpperCamelCase : Dict = None , ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) SCREAMING_SNAKE_CASE = self.encoder( __lowercase , output_hidden_states=__lowercase , return_dict=__lowercase , ) SCREAMING_SNAKE_CASE = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=__lowercase , hidden_states=encoder_outputs.hidden_states , ) class lowercase ( nn.Module ): def __init__( self : Tuple , _UpperCamelCase : Optional[Any] ) -> List[Any]: '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE = nn.Linear(config.hidden_size , config.hidden_size ) def __snake_case( self : Union[str, Any] , _UpperCamelCase : Union[str, Any] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = self.dense(__lowercase ) return output @add_start_docstrings( """\n PoolFormer Model transformer with an image classification head on top\n """ , __A , ) class lowercase ( __A ): def __init__( self : Tuple , _UpperCamelCase : int ) -> Optional[int]: '''simple docstring''' super().__init__(__lowercase ) SCREAMING_SNAKE_CASE = config.num_labels SCREAMING_SNAKE_CASE = PoolFormerModel(__lowercase ) # Final norm SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head SCREAMING_SNAKE_CASE = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__lowercase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__lowercase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __snake_case( self : Optional[Any] , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : Optional[Any] = None , _UpperCamelCase : Any = None , _UpperCamelCase : List[str] = None , ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict SCREAMING_SNAKE_CASE = self.poolformer( __lowercase , output_hidden_states=__lowercase , return_dict=__lowercase , ) SCREAMING_SNAKE_CASE = outputs[0] SCREAMING_SNAKE_CASE = self.classifier(self.norm(__lowercase ).mean([-2, -1] ) ) SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: SCREAMING_SNAKE_CASE = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): SCREAMING_SNAKE_CASE = "single_label_classification" else: SCREAMING_SNAKE_CASE = "multi_label_classification" if self.config.problem_type == "regression": SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: SCREAMING_SNAKE_CASE = loss_fct(__lowercase , __lowercase ) elif self.config.problem_type == "single_label_classification": SCREAMING_SNAKE_CASE = CrossEntropyLoss() SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() SCREAMING_SNAKE_CASE = loss_fct(__lowercase , __lowercase ) if not return_dict: SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=__lowercase , logits=__lowercase , hidden_states=outputs.hidden_states )
359
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCamelCase : str = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} _lowerCamelCase : int = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowerCamelCase : List[Any] = { '''allenai/led-base-16384''': 1_63_84, } class lowercase ( a ): lowercase__ : Tuple = VOCAB_FILES_NAMES lowercase__ : Any = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Optional[Any] = LEDTokenizer lowercase__ : str = ["""input_ids""", """attention_mask"""] def __init__( self : Union[str, Any] , _UpperCamelCase : Tuple=None , _UpperCamelCase : str=None , _UpperCamelCase : Tuple=None , _UpperCamelCase : List[str]="replace" , _UpperCamelCase : str="<s>" , _UpperCamelCase : List[Any]="</s>" , _UpperCamelCase : List[Any]="</s>" , _UpperCamelCase : List[str]="<s>" , _UpperCamelCase : Tuple="<unk>" , _UpperCamelCase : List[Any]="<pad>" , _UpperCamelCase : Tuple="<mask>" , _UpperCamelCase : List[str]=False , _UpperCamelCase : List[Any]=True , **_UpperCamelCase : Optional[Any] , ) -> Tuple: '''simple docstring''' super().__init__( _UpperCamelCase , _UpperCamelCase , tokenizer_file=_UpperCamelCase , errors=_UpperCamelCase , bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , sep_token=_UpperCamelCase , cls_token=_UpperCamelCase , unk_token=_UpperCamelCase , pad_token=_UpperCamelCase , mask_token=_UpperCamelCase , add_prefix_space=_UpperCamelCase , trim_offsets=_UpperCamelCase , **_UpperCamelCase , ) SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , _UpperCamelCase ) != add_prefix_space: SCREAMING_SNAKE_CASE = getattr(_UpperCamelCase , pre_tok_state.pop("type" ) ) SCREAMING_SNAKE_CASE = add_prefix_space SCREAMING_SNAKE_CASE = pre_tok_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` SCREAMING_SNAKE_CASE = "post_processor" SCREAMING_SNAKE_CASE = getattr(self.backend_tokenizer , _UpperCamelCase , _UpperCamelCase ) if tokenizer_component_instance: SCREAMING_SNAKE_CASE = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: SCREAMING_SNAKE_CASE = tuple(state["sep"] ) if "cls" in state: SCREAMING_SNAKE_CASE = tuple(state["cls"] ) SCREAMING_SNAKE_CASE = False if state.get("add_prefix_space" , _UpperCamelCase ) != add_prefix_space: SCREAMING_SNAKE_CASE = add_prefix_space SCREAMING_SNAKE_CASE = True if state.get("trim_offsets" , _UpperCamelCase ) != trim_offsets: SCREAMING_SNAKE_CASE = trim_offsets SCREAMING_SNAKE_CASE = True if changes_to_apply: SCREAMING_SNAKE_CASE = getattr(_UpperCamelCase , state.pop("type" ) ) SCREAMING_SNAKE_CASE = component_class(**_UpperCamelCase ) setattr(self.backend_tokenizer , _UpperCamelCase , _UpperCamelCase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def __snake_case( self : int ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def __snake_case( self : Optional[int] , _UpperCamelCase : Optional[int] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else value SCREAMING_SNAKE_CASE = value def __snake_case( self : List[Any] , *_UpperCamelCase : Optional[int] , **_UpperCamelCase : Union[str, Any] ) -> BatchEncoding: '''simple docstring''' SCREAMING_SNAKE_CASE = kwargs.get("is_split_into_words" , _UpperCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : int , *_UpperCamelCase : Dict , **_UpperCamelCase : Tuple ) -> BatchEncoding: '''simple docstring''' SCREAMING_SNAKE_CASE = kwargs.get("is_split_into_words" , _UpperCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : List[Any] , _UpperCamelCase : str , _UpperCamelCase : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = self._tokenizer.model.save(_UpperCamelCase , name=_UpperCamelCase ) return tuple(_UpperCamelCase ) def __snake_case( self : List[Any] , _UpperCamelCase : List[Any] , _UpperCamelCase : int=None ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __snake_case( self : Dict , _UpperCamelCase : List[int] , _UpperCamelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.sep_token_id] SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __snake_case( self : Optional[Any] , _UpperCamelCase : Union[Dict[str, EncodedInput], BatchEncoding] , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : Optional[bool] = None , ) -> dict: '''simple docstring''' SCREAMING_SNAKE_CASE = super()._pad( encoded_inputs=_UpperCamelCase , max_length=_UpperCamelCase , padding_strategy=_UpperCamelCase , pad_to_multiple_of=_UpperCamelCase , return_attention_mask=_UpperCamelCase , ) # Load from model defaults if return_attention_mask is None: SCREAMING_SNAKE_CASE = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: SCREAMING_SNAKE_CASE = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. SCREAMING_SNAKE_CASE = len(encoded_inputs["global_attention_mask"] ) != len(_UpperCamelCase ) if needs_to_be_padded: SCREAMING_SNAKE_CASE = len(_UpperCamelCase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` SCREAMING_SNAKE_CASE = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": SCREAMING_SNAKE_CASE = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
206
0
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _lowerCAmelCase : Tuple = logging.get_logger(__name__) _lowerCAmelCase : List[str] = {'''vocab_file''': '''vocab.txt''', '''emoji_file''': '''emoji.json'''} _lowerCAmelCase : Any = { '''vocab_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt''', }, '''emoji_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json''', }, } _lowerCAmelCase : List[Any] = { '''abeja/gpt-neox-japanese-2.7b''': 2_048, } def __snake_case ( _lowerCAmelCase : Any , _lowerCAmelCase : str ) -> Union[str, Any]: with open(_lowerCAmelCase , "r" , encoding="utf-8" ) as f: A_ : Tuple = json.loads(f.read() ) A_ : Any = collections.OrderedDict() A_ : Optional[Any] = collections.OrderedDict() A_ : Dict = collections.OrderedDict() with open(_lowerCAmelCase , "r" , encoding="utf-8" ) as f: A_ : List[str] = f.readlines() A_ : int = [[t.rstrip("\n" )] if (t == "," or "," not in t) else t.rstrip("\n" ).split("," ) for t in token] for idx, b in enumerate(_lowerCAmelCase ): A_ : Tuple = b A_ : Optional[int] = idx for wd in b: A_ : Optional[Any] = idx return vocab, raw_vocab, ids_to_tokens, emoji class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self :List[Any] , snake_case :List[Any] , snake_case :List[str] , snake_case :List[Any]="<|endoftext|>" , snake_case :Dict="<|endoftext|>" , snake_case :List[Any]="<|startoftext|>" , snake_case :int="<|endoftext|>" , snake_case :Dict=False , **snake_case :List[Any] , ): '''simple docstring''' super().__init__( unk_token=snake_case , pad_token=snake_case , bos_token=snake_case , eos_token=snake_case , do_clean_text=snake_case , **snake_case , ) if not os.path.isfile(snake_case ): raise ValueError( f"Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained" " model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) if not os.path.isfile(snake_case ): raise ValueError( f"Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google" " pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) A_ : List[str] = do_clean_text A_ , A_ , A_ , A_ : List[Any] = load_vocab_and_emoji(snake_case , snake_case ) A_ : Any = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def SCREAMING_SNAKE_CASE ( self :Union[str, Any] ): '''simple docstring''' return len(self.raw_vocab ) def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' return dict(self.raw_vocab , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , snake_case :Tuple ): '''simple docstring''' return self.subword_tokenizer.tokenize(snake_case , clean=self.do_clean_text ) def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :List[str] ): '''simple docstring''' return self.vocab.get(snake_case , self.vocab.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , snake_case :Any ): '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :int ): '''simple docstring''' A_ : Tuple = "".join(snake_case ).strip() return out_string def SCREAMING_SNAKE_CASE ( self :str , snake_case :"Conversation" ): '''simple docstring''' A_ : List[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(snake_case , add_special_tokens=snake_case ) + [self.eos_token_id] ) if len(snake_case ) > self.model_max_length: A_ : Any = input_ids[-self.model_max_length :] return input_ids def SCREAMING_SNAKE_CASE ( self :Any , snake_case :str , snake_case :Optional[str] = None ): '''simple docstring''' A_ : int = 0 if os.path.isdir(snake_case ): A_ : Optional[int] = os.path.join( snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) A_ : Dict = os.path.join( snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["emoji_file"] ) else: A_ : Any = ( (filename_prefix + "-" if filename_prefix else "") + save_directory + VOCAB_FILES_NAMES["vocab_file"] ) A_ : Optional[Any] = ( (filename_prefix + "-" if filename_prefix else "") + save_directory + VOCAB_FILES_NAMES["emoji_file"] ) with open(snake_case , "w" , encoding="utf-8" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f"Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive." " Please check that the vocabulary is not corrupted!" ) A_ : Any = token_index writer.write(",".join(snake_case ) + "\n" ) index += 1 with open(snake_case , "w" , encoding="utf-8" ) as writer: json.dump(self.emoji , snake_case ) return vocab_file, emoji_file class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" def __init__( self :List[str] , snake_case :str , snake_case :List[str] , snake_case :List[str] ): '''simple docstring''' A_ : int = vocab # same as swe A_ : int = ids_to_tokens # same as bpe A_ : Union[str, Any] = emoji A_ : Optional[int] = np.max([len(snake_case ) for w in self.vocab.keys()] ) A_ : Union[str, Any] = re.compile(R"(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)" ) A_ : Union[str, Any] = re.compile(R"[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*" ) A_ : Optional[Any] = re.compile(R"[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}" ) A_ : List[Any] = re.compile( R"([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) A_ : str = re.compile( R"(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) A_ : str = re.compile( R"((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*" ) A_ : List[str] = "─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿" A_ : Union[str, Any] = "▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟" A_ : List[str] = str.maketrans({k: "<BLOCK>" for k in keisen + blocks} ) def __len__( self :Tuple ): '''simple docstring''' return len(self.ids_to_tokens ) def SCREAMING_SNAKE_CASE ( self :List[Any] , snake_case :int ): '''simple docstring''' A_ : Union[str, Any] = self.content_repattera.sub("<URL>" , snake_case ) A_ : List[Any] = self.content_repattera.sub("<EMAIL>" , snake_case ) A_ : List[str] = self.content_repattera.sub("<TEL>" , snake_case ) A_ : List[str] = self.content_repattera.sub("<DATE>" , snake_case ) A_ : Dict = self.content_repattera.sub("<DATE>" , snake_case ) A_ : str = self.content_repattera.sub("<PRICE>" , snake_case ) A_ : Tuple = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: A_ : Optional[Any] = content.replace("<BLOCK><BLOCK>" , "<BLOCK>" ) return content def SCREAMING_SNAKE_CASE ( self :Any , snake_case :int , snake_case :int=False ): '''simple docstring''' A_ : List[str] = text.replace(" " , "<SP>" ) A_ : Optional[int] = text.replace(" " , "<SP>" ) A_ : Optional[int] = text.replace("\r\n" , "<BR>" ) A_ : Dict = text.replace("\n" , "<BR>" ) A_ : List[str] = text.replace("\r" , "<BR>" ) A_ : List[Any] = text.replace("\t" , "<TAB>" ) A_ : List[str] = text.replace("—" , "ー" ) A_ : Union[str, Any] = text.replace("−" , "ー" ) for k, v in self.emoji["emoji"].items(): if k in text: A_ : int = text.replace(snake_case , snake_case ) if clean: A_ : Optional[Any] = self.clean_text(snake_case ) def check_simbol(snake_case :Dict ): A_ : Optional[int] = x.encode() if len(snake_case ) == 1 and len(snake_case ) == 2: A_ : Optional[Any] = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xc2_a1 and c <= 0xc2_bf) or (c >= 0xc7_80 and c <= 0xc7_83) or (c >= 0xca_b9 and c <= 0xcb_bf) or (c >= 0xcc_80 and c <= 0xcd_a2) ): return True return False def checkuae(snake_case :Dict ): A_ : Dict = x.encode() if len(snake_case ) == 1 and len(snake_case ) == 3: A_ : List[Any] = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xe2_80_80 and c <= 0xe2_b0_7f: return True return False A_ : Dict = 0 A_ : List[str] = [] while pos < len(snake_case ): A_ : str = min(len(snake_case ) , pos + self.maxlen + 1 ) if text[pos] == "<" else pos + 3 A_ : Tuple = [] # (token_id, token, pos) for e in range(snake_case , snake_case , -1 ): A_ : Any = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(snake_case ) > 2: A_ : Optional[int] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(snake_case ) > 0: # the smallest token_id is adopted A_ , A_ , A_ : Any = sorted(snake_case , key=lambda snake_case : x[0] )[0] result.append(snake_case ) A_ : Optional[Any] = e else: A_ : Union[str, Any] = pos + 1 A_ : List[Any] = text[pos:end] if check_simbol(snake_case ): result.append("<KIGOU>" ) elif checkuae(snake_case ): result.append("<U2000U2BFF>" ) else: for i in wd.encode("utf-8" ): result.append("<|byte%d|>" % i ) A_ : Dict = end return result def SCREAMING_SNAKE_CASE ( self :Optional[int] , snake_case :Union[str, Any] , snake_case :Dict="\n" ): '''simple docstring''' A_ : Tuple = [] A_ : Optional[int] = [] A_ : Union[str, Any] = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(snake_case ) > 0: words.append(bytearray(snake_case ).decode("utf-8" , errors="replace" ) ) A_ : str = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["emoji_inv"][word] ) elif word == "<SP>": words.append(" " ) elif word == "<BR>": words.append(snake_case ) elif word == "<TAB>": words.append("\t" ) elif word == "<BLOCK>": words.append("▀" ) elif word == "<KIGOU>": words.append("ǀ" ) elif word == "<U2000U2BFF>": words.append("‖" ) else: words.append(snake_case ) if len(snake_case ) > 0: words.append(bytearray(snake_case ).decode("utf-8" , errors="replace" ) ) A_ : Dict = "".join(snake_case ) return text
300
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class __magic_name__ : """simple docstring""" def __init__( self :Tuple , snake_case :Optional[Any] , snake_case :Tuple=13 , snake_case :Dict=7 , snake_case :List[Any]=True , snake_case :List[Any]=True , snake_case :Dict=True , snake_case :Any=True , snake_case :Optional[int]=99 , snake_case :Any=32 , snake_case :Dict=2 , snake_case :int=4 , snake_case :Optional[int]=37 , snake_case :List[str]="gelu" , snake_case :List[Any]=0.1 , snake_case :Optional[Any]=0.1 , snake_case :Tuple=512 , snake_case :Tuple=16 , snake_case :Tuple=2 , snake_case :Optional[int]=0.02 , snake_case :str=3 , snake_case :Optional[int]=4 , snake_case :List[str]=None , snake_case :Tuple=1_000 , ): '''simple docstring''' A_ : str = parent A_ : str = batch_size A_ : str = seq_length A_ : Any = is_training A_ : Any = use_input_mask A_ : str = use_token_type_ids A_ : Tuple = use_labels A_ : Optional[Any] = vocab_size A_ : Dict = hidden_size A_ : str = num_hidden_layers A_ : Dict = num_attention_heads A_ : str = intermediate_size A_ : int = hidden_act A_ : List[Any] = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Optional[Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : Any = type_sequence_label_size A_ : Dict = initializer_range A_ : Any = num_labels A_ : Optional[int] = num_choices A_ : Optional[Any] = scope A_ : Any = range_bbox def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment A_ : Tuple = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: A_ : str = bbox[i, j, 3] A_ : Union[str, Any] = bbox[i, j, 1] A_ : List[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: A_ : Any = bbox[i, j, 2] A_ : Tuple = bbox[i, j, 0] A_ : int = t A_ : int = tf.convert_to_tensor(snake_case ) A_ : Any = None if self.use_input_mask: A_ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) A_ : str = None if self.use_token_type_ids: A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Dict = None A_ : List[Any] = None A_ : List[str] = None if self.use_labels: A_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : str = ids_tensor([self.batch_size] , self.num_choices ) A_ : int = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self :str , snake_case :Dict , snake_case :Union[str, Any] , snake_case :int , snake_case :int , snake_case :Union[str, Any] , snake_case :Tuple , snake_case :Optional[int] , snake_case :List[Any] ): '''simple docstring''' A_ : Any = TFLayoutLMModel(config=snake_case ) A_ : Tuple = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A_ : str = model(snake_case , snake_case , token_type_ids=snake_case ) A_ : List[Any] = model(snake_case , snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self :Optional[int] , snake_case :Any , snake_case :List[Any] , snake_case :List[str] , snake_case :Optional[Any] , snake_case :Dict , snake_case :Any , snake_case :Union[str, Any] , snake_case :List[Any] ): '''simple docstring''' A_ : Optional[int] = TFLayoutLMForMaskedLM(config=snake_case ) A_ : Tuple = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :Dict , snake_case :Tuple , snake_case :Tuple , snake_case :List[str] , snake_case :Tuple , snake_case :str , snake_case :Optional[int] , snake_case :Any ): '''simple docstring''' A_ : Union[str, Any] = self.num_labels A_ : int = TFLayoutLMForSequenceClassification(config=snake_case ) A_ : Optional[int] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :Dict , snake_case :str , snake_case :Optional[Any] , snake_case :int , snake_case :Any , snake_case :Tuple , snake_case :List[str] , snake_case :Union[str, Any] ): '''simple docstring''' A_ : List[Any] = self.num_labels A_ : str = TFLayoutLMForTokenClassification(config=snake_case ) A_ : Union[str, Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self :int , snake_case :List[str] , snake_case :Optional[int] , snake_case :Union[str, Any] , snake_case :List[Any] , snake_case :int , snake_case :Any , snake_case :Union[str, Any] , snake_case :Any ): '''simple docstring''' A_ : Optional[Any] = TFLayoutLMForQuestionAnswering(config=snake_case ) A_ : List[Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : int = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Optional[Any] = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class __magic_name__ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __UpperCamelCase = ( { '''feature-extraction''': TFLayoutLMModel, '''fill-mask''': TFLayoutLMForMaskedLM, '''text-classification''': TFLayoutLMForSequenceClassification, '''token-classification''': TFLayoutLMForTokenClassification, '''zero-shot''': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __UpperCamelCase = False __UpperCamelCase = True __UpperCamelCase = 10 def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : Tuple = TFLayoutLMModelTester(self ) A_ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : List[str] = TFLayoutLMModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip("Onnx compliancy broke with TF 2.10" ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' pass def __snake_case ( ) -> Optional[Any]: # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off A_ : int = tf.convert_to_tensor([[101,1019,1014,1016,1037,12849,4747,1004,14246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,11300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,19274,2772,6205,27814,16147,16147,4343,2047,10283,10969,14389,1012,2338,102]] ) # noqa: E231 A_ : int = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 A_ : Union[str, Any] = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 A_ : List[Any] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) A_ : Tuple = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class __magic_name__ ( unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : str = TFLayoutLMModel.from_pretrained("microsoft/layoutlm-base-uncased" ) A_ , A_ , A_ , A_ , A_ : Tuple = prepare_layoutlm_batch_inputs() # forward pass A_ : Tuple = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the sequence output on [0, :3, :3] A_ : List[Any] = tf.convert_to_tensor( [[0.1785, -0.1947, -0.0425], [-0.3254, -0.2807, 0.2553], [-0.5391, -0.3322, 0.3364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1e-3 ) ) # test the pooled output on [1, :3] A_ : Optional[Any] = tf.convert_to_tensor([-0.6580, -0.0214, 0.8552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , snake_case , atol=1e-3 ) ) @slow def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : Union[str, Any] = TFLayoutLMForSequenceClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=2 ) A_ , A_ , A_ , A_ , A_ : Any = prepare_layoutlm_batch_inputs() # forward pass A_ : Dict = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar A_ : List[str] = outputs.loss A_ : Union[str, Any] = (2,) self.assertEqual(loss.shape , snake_case ) # test the shape of the logits A_ : Tuple = outputs.logits A_ : Tuple = (2, 2) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : int = TFLayoutLMForTokenClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=13 ) A_ , A_ , A_ , A_ , A_ : Optional[int] = prepare_layoutlm_batch_inputs() # forward pass A_ : Union[str, Any] = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) # test the shape of the logits A_ : Dict = outputs.logits A_ : List[Any] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : Optional[Any] = TFLayoutLMForQuestionAnswering.from_pretrained("microsoft/layoutlm-base-uncased" ) A_ , A_ , A_ , A_ , A_ : str = prepare_layoutlm_batch_inputs() # forward pass A_ : Union[str, Any] = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the shape of the logits A_ : Union[str, Any] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , snake_case ) self.assertEqual(outputs.end_logits.shape , snake_case )
300
1
"""simple docstring""" def _A ( UpperCamelCase_ : Optional[Any]) -> Optional[int]: '''simple docstring''' __lowercase = current_set.copy() for row_index, row in enumerate(UpperCamelCase_): __lowercase = row[0] for column_index, column in enumerate(UpperCamelCase_): if magnitude == 0: __lowercase = column continue __lowercase = column / magnitude # Subtract to cancel term __lowercase = current_set[0] __lowercase = [first_row] __lowercase = current_set[1::] for row in current_set: __lowercase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(UpperCamelCase_) continue for column_index in range(len(UpperCamelCase_)): temp_row.append(first_row[column_index] - row[column_index]) final_set.append(UpperCamelCase_) # Create next recursion iteration set if len(final_set[0]) != 3: __lowercase = final_set[0] __lowercase = [] __lowercase = [] for row in final_set[1::]: current_first_column.append(row[0]) next_iteration.append(row[1::]) __lowercase = simplify(UpperCamelCase_) for i in range(len(UpperCamelCase_)): resultant[i].insert(0, current_first_column[i]) resultant.insert(0, UpperCamelCase_) __lowercase = resultant return final_set def _A ( UpperCamelCase_ : Any) -> List[Any]: '''simple docstring''' if len(UpperCamelCase_) == 0: raise IndexError("solve_simultaneous() requires n lists of length n+1") __lowercase = len(UpperCamelCase_) + 1 if any(len(UpperCamelCase_) != _length for item in equations): raise IndexError("solve_simultaneous() requires n lists of length n+1") for row in equations: if any(not isinstance(UpperCamelCase_, (int, float)) for column in row): raise ValueError("solve_simultaneous() requires lists of integers") if len(UpperCamelCase_) == 1: return [equations[0][-1] / equations[0][0]] __lowercase = equations.copy() if any(0 in row for row in data_set): __lowercase = data_set.copy() __lowercase = [] for row_index, row in enumerate(UpperCamelCase_): if 0 not in row: __lowercase = data_set.pop(UpperCamelCase_) break if not full_row: raise ValueError("solve_simultaneous() requires at least 1 full equation") data_set.insert(0, UpperCamelCase_) __lowercase = data_set.copy() __lowercase = simplify(UpperCamelCase_) __lowercase = simplified[::-1] __lowercase = [] for row in simplified: __lowercase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0) continue solutions.append(current_solution / row[-2]) continue __lowercase = row.copy()[: len(UpperCamelCase_) - 1 :] while temp_row[0] == 0: temp_row.pop(0) if len(UpperCamelCase_) == 0: solutions.append(0) continue __lowercase = temp_row[1::] __lowercase = temp_row[::-1] for column_index, column in enumerate(UpperCamelCase_): current_solution -= column * solutions[column_index] solutions.append(UpperCamelCase_) __lowercase = [] for item in solutions: final.append(float(round(UpperCamelCase_, 5))) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() _a = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
355
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration _a = 5_00_00 _a = 50_00 _a , _a = os.path.split(__file__) _a = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : List[str]) -> List[str]: '''simple docstring''' for i in range(UpperCamelCase_): __lowercase = dataset[i] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : List[Any], UpperCamelCase_ : int) -> Dict: '''simple docstring''' for i in range(0, len(UpperCamelCase_), UpperCamelCase_): __lowercase = dataset[i : i + batch_size] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : Any, UpperCamelCase_ : Optional[int]) -> List[str]: '''simple docstring''' with dataset.formatted_as(type=UpperCamelCase_): for i in range(UpperCamelCase_): __lowercase = dataset[i] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : str, UpperCamelCase_ : Union[str, Any], UpperCamelCase_ : Union[str, Any]) -> Dict: '''simple docstring''' with dataset.formatted_as(type=UpperCamelCase_): for i in range(0, UpperCamelCase_, UpperCamelCase_): __lowercase = dataset[i : i + batch_size] def _A ( ) -> List[str]: '''simple docstring''' __lowercase = {"num examples": SPEED_TEST_N_EXAMPLES} __lowercase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] __lowercase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset") __lowercase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32")), "numbers": datasets.Value("float32")}) __lowercase = generate_example_dataset( os.path.join(UpperCamelCase_, "dataset.arrow"), UpperCamelCase_, num_examples=UpperCamelCase_, seq_shapes={"list": (100,)}, ) print("first set of iterations") for func, kwargs in functions: print(func.__name__, str(UpperCamelCase_)) __lowercase = func(UpperCamelCase_, **UpperCamelCase_) print("shuffling dataset") __lowercase = dataset.shuffle() print("Second set of iterations (after shuffling") for func, kwargs in functions_shuffled: print("shuffled ", func.__name__, str(UpperCamelCase_)) __lowercase = func( UpperCamelCase_, **UpperCamelCase_) with open(UpperCamelCase_, "wb") as f: f.write(json.dumps(UpperCamelCase_).encode("utf-8")) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
144
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowerCamelCase_ ( ): lowerCamelCase_ = ArgumentParser( description=( "PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=lowerCamelCase__ , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=lowerCamelCase__ , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=lowerCamelCase__ ) return parser.parse_args() def lowerCamelCase_ ( ): lowerCamelCase_ = parse_args() # Import training_script as a module. lowerCamelCase_ = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) lowerCamelCase_ = script_fpath.stem lowerCamelCase_ = importlib.import_module(lowerCamelCase__ ) # Patch sys.argv lowerCamelCase_ = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
19
import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging __A =logging.get_logger(__name__) def lowerCamelCase_ ( ): # Get the sagemaker specific mp parameters from smp_options variable. lowerCamelCase_ = os.getenv("SM_HP_MP_PARAMETERS" , "{}" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. lowerCamelCase_ = json.loads(lowerCamelCase__ ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. lowerCamelCase_ = os.getenv("SM_FRAMEWORK_PARAMS" , "{}" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". lowerCamelCase_ = json.loads(lowerCamelCase__ ) if not mpi_options.get("sagemaker_mpi_enabled" , lowerCamelCase__ ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("smdistributed" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = field( default='' , metadata={'help': 'Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'} , ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: super().__post_init__() warnings.warn( "`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use " "`TrainingArguments` instead." , lowercase , ) @cached_property def SCREAMING_SNAKE_CASE_( self ) -> "torch.device": logger.info("PyTorch: setting up devices" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( "torch.distributed process group is initialized, but local_rank == -1. " "In order to use Torch DDP, launch your script with `python -m torch.distributed.launch" ) if self.no_cuda: lowerCamelCase_ = torch.device("cpu" ) lowerCamelCase_ = 0 elif is_sagemaker_model_parallel_available(): lowerCamelCase_ = smp.local_rank() lowerCamelCase_ = torch.device("cuda" , lowercase ) lowerCamelCase_ = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="smddp" , timeout=self.ddp_timeout_delta ) lowerCamelCase_ = int(os.getenv("SMDATAPARALLEL_LOCAL_RANK" ) ) lowerCamelCase_ = torch.device("cuda" , self.local_rank ) lowerCamelCase_ = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 lowerCamelCase_ = torch.device("cuda:0" if torch.cuda.is_available() else "cpu" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. lowerCamelCase_ = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="nccl" , timeout=self.ddp_timeout_delta ) lowerCamelCase_ = torch.device("cuda" , self.local_rank ) lowerCamelCase_ = 1 if device.type == "cuda": torch.cuda.set_device(lowercase ) return device @property def SCREAMING_SNAKE_CASE_( self ) -> Tuple: if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def SCREAMING_SNAKE_CASE_( self ) -> List[str]: return not is_sagemaker_model_parallel_available() @property def SCREAMING_SNAKE_CASE_( self ) -> Dict: return False
19
1
def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(__a ) == 0: raise ValueError('Input list must be a non empty list' ) if len(__a ) == 1: return True snake_case_ : Union[str, Any] = series[1] - series[0] for index in range(len(__a ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(__a ) == 0: raise ValueError('Input list must be a non empty list' ) snake_case_ : List[str] = 0 for val in series: answer += val return answer / len(__a ) if __name__ == "__main__": import doctest doctest.testmod()
88
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys _SCREAMING_SNAKE_CASE = subprocess.check_output("""git merge-base main HEAD""".split()).decode("""utf-8""") _SCREAMING_SNAKE_CASE = ( subprocess.check_output(F'''git diff --diff-filter=d --name-only {fork_point_sha}'''.split()).decode("""utf-8""").split() ) _SCREAMING_SNAKE_CASE = """|""".join(sys.argv[1:]) _SCREAMING_SNAKE_CASE = re.compile(RF'''^({joined_dirs}).*?\.py$''') _SCREAMING_SNAKE_CASE = [x for x in modified_files if regex.match(x)] print(""" """.join(relevant_modified_files), end="""""")
88
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor __lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase__( __A ): def __init__( self ,*__UpperCAmelCase ,**__UpperCAmelCase ) -> None: warnings.warn( 'The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use SegformerImageProcessor instead.' ,__UpperCAmelCase ,) super().__init__(*__UpperCAmelCase ,**__UpperCAmelCase )
221
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class UpperCamelCase__: def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> int: return None class UpperCamelCase__: def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> Tuple: return None class UpperCamelCase__( unittest.TestCase ): lowerCAmelCase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ) -> int: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__UpperCAmelCase ,'tf' ,12 ,**__UpperCAmelCase ) @require_torch @slow def snake_case__ ( self ) -> int: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__UpperCAmelCase ,'pt' ,12 ,**__UpperCAmelCase ) @require_torch @slow def snake_case__ ( self ) -> Optional[Any]: from transformers import BertModel A__ = ['[UNK]', '[SEP]', '[CLS]', '[PAD]', '[MASK]', 'some', 'other', 'words'] with NamedTemporaryFile(mode='w+t' ) as vocab_file: vocab_file.write('\n'.join(__UpperCAmelCase ) ) vocab_file.flush() A__ = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: A__ = BertModel(BertConfig(vocab_size=len(__UpperCAmelCase ) ) ) model.save_pretrained(__UpperCAmelCase ) self._test_export(__UpperCAmelCase ,'pt' ,12 ,__UpperCAmelCase ) @require_tf @slow def snake_case__ ( self ) -> Optional[Any]: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: A__ = self._test_export(__UpperCAmelCase ,'tf' ,12 ,**__UpperCAmelCase ) A__ = quantize(Path(__UpperCAmelCase ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__UpperCAmelCase ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) @require_torch @slow def snake_case__ ( self ) -> str: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: A__ = self._test_export(__UpperCAmelCase ,'pt' ,12 ,**__UpperCAmelCase ) A__ = quantize(__UpperCAmelCase ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__UpperCAmelCase ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase=None ,**__UpperCAmelCase ) -> Union[str, Any]: try: # Compute path with TemporaryDirectory() as tempdir: A__ = Path(__UpperCAmelCase ).joinpath('model.onnx' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,**__UpperCAmelCase ) return path except Exception as e: self.fail(__UpperCAmelCase ) @require_torch @require_tokenizers @slow def snake_case__ ( self ) -> Optional[Any]: from transformers import BertModel A__ = BertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) A__ = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(__UpperCAmelCase ,__UpperCAmelCase ,'pt' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ) -> Optional[Any]: from transformers import TFBertModel A__ = TFBertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) A__ = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(__UpperCAmelCase ,__UpperCAmelCase ,'tf' ) def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> str: A__ = FeatureExtractionPipeline(__UpperCAmelCase ,__UpperCAmelCase ) A__ = ['input_ids', 'token_type_ids', 'attention_mask', 'output_0', 'output_1'] A__ , A__ , A__ , A__ = infer_shapes(__UpperCAmelCase ,__UpperCAmelCase ) # Assert all variables are present self.assertEqual(len(__UpperCAmelCase ) ,len(__UpperCAmelCase ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] ,__UpperCAmelCase ) self.assertSequenceEqual(variable_names[3:] ,__UpperCAmelCase ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] ,{0: 'batch', 1: 'sequence'} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['output_0'] ,{0: 'batch', 1: 'sequence'} ) self.assertDictEqual(shapes['output_1'] ,{0: 'batch'} ) def snake_case__ ( self ) -> Union[str, Any]: A__ = ['input_ids', 'attention_mask', 'token_type_ids'] A__ = {'input_ids': [1, 2, 3, 4], 'attention_mask': [0, 0, 0, 0], 'token_type_ids': [1, 1, 1, 1]} A__ , A__ = ensure_valid_input(FuncContiguousArgs() ,__UpperCAmelCase ,__UpperCAmelCase ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(__UpperCAmelCase ) ,3 ) # Should have exactly the same input names self.assertEqual(set(__UpperCAmelCase ) ,set(__UpperCAmelCase ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(__UpperCAmelCase ,(tokens['input_ids'], tokens['token_type_ids'], tokens['attention_mask']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) A__ , A__ = ensure_valid_input(FuncNonContiguousArgs() ,__UpperCAmelCase ,__UpperCAmelCase ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(__UpperCAmelCase ) ,1 ) self.assertEqual(len(__UpperCAmelCase ) ,1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] ,tokens['input_ids'] ) self.assertEqual(ordered_input_names[0] ,'input_ids' ) def snake_case__ ( self ) -> Union[str, Any]: A__ = generate_identified_filename(Path('/home/something/my_fake_model.onnx' ) ,'-test' ) self.assertEqual('/home/something/my_fake_model-test.onnx' ,generated.as_posix() )
221
1
from math import sqrt def __lowerCamelCase ( UpperCAmelCase_ : int ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" a :str = True # 0 and 1 are none primes. if number <= 1: a :Union[str, Any] = False for divisor in range(2 , int(round(sqrt(UpperCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: a :List[str] = False break # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'status' must been from type bool" return status def __lowerCamelCase ( UpperCAmelCase_ : List[str] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N a :Optional[int] = list(range(2 , n + 1 ) ) a :Union[str, Any] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(UpperCAmelCase_ ) ): for j in range(i + 1 , len(UpperCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): a :Optional[Any] = 0 # filters actual prime numbers. a :int = [x for x in begin_list if x != 0] # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'ans' must been from type list" return ans def __lowerCamelCase ( UpperCAmelCase_ : Dict ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" a :List[str] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(UpperCAmelCase_ ): ans.append(UpperCAmelCase_ ) # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'ans' must been from type list" return ans def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" a :str = [] # this list will be returns of the function. # potential prime number factors. a :Any = 2 a :int = number if number == 0 or number == 1: ans.append(UpperCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(UpperCAmelCase_ ): while quotient != 1: if is_prime(UpperCAmelCase_ ) and (quotient % factor == 0): ans.append(UpperCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(UpperCAmelCase_ ) # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'ans' must been from type list" return ans def __lowerCamelCase ( UpperCAmelCase_ : List[Any] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" a :Tuple = 0 # prime factorization of 'number' a :Union[str, Any] = prime_factorization(UpperCAmelCase_ ) a :Tuple = max(UpperCAmelCase_ ) # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'ans' must been from type int" return ans def __lowerCamelCase ( UpperCAmelCase_ : Any ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" a :Any = 0 # prime factorization of 'number' a :Any = prime_factorization(UpperCAmelCase_ ) a :Any = min(UpperCAmelCase_ ) # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'ans' must been from type int" return ans def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , UpperCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def __lowerCamelCase ( UpperCAmelCase_ : Optional[int] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , UpperCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def __lowerCamelCase ( UpperCAmelCase_ : List[str] ): """simple docstring""" assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (number > 2) and is_even(UpperCAmelCase_ ) ), "'number' must been an int, even and > 2" a :Optional[int] = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' a :Optional[Any] = get_prime_numbers(UpperCAmelCase_ ) a :List[Any] = len(UpperCAmelCase_ ) # run variable for while-loops. a :int = 0 a :Optional[int] = None # exit variable. for break up the loops a :Optional[Any] = True while i < len_pn and loop: a :List[Any] = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: a :Optional[int] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (len(UpperCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def __lowerCamelCase ( UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple ): """simple docstring""" assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." a :Dict = 0 while numbera != 0: a :Any = numbera % numbera a :str = numbera a :Dict = rest # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." a :List[str] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' a :Tuple = prime_factorization(UpperCAmelCase_ ) a :Tuple = prime_factorization(UpperCAmelCase_ ) elif numbera == 1 or numbera == 1: a :str = [] a :Any = [] a :List[Any] = max(UpperCAmelCase_ , UpperCAmelCase_ ) a :List[Any] = 0 a :Optional[int] = 0 a :Optional[int] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: a :Union[str, Any] = prime_fac_a.count(UpperCAmelCase_ ) a :int = prime_fac_a.count(UpperCAmelCase_ ) for _ in range(max(UpperCAmelCase_ , UpperCAmelCase_ ) ): ans *= n else: a :Optional[Any] = prime_fac_a.count(UpperCAmelCase_ ) for _ in range(UpperCAmelCase_ ): ans *= n done.append(UpperCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: a :Dict = prime_fac_a.count(UpperCAmelCase_ ) for _ in range(UpperCAmelCase_ ): ans *= n done.append(UpperCAmelCase_ ) # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def __lowerCamelCase ( UpperCAmelCase_ : Optional[int] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n >= 0), "'number' must been a positive int" a :int = 0 a :Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(UpperCAmelCase_ ): ans += 1 # precondition assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and is_prime( UpperCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] ): """simple docstring""" assert ( is_prime(UpperCAmelCase_ ) and is_prime(UpperCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" a :Any = p_number_a + 1 # jump to the next number a :Optional[Any] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(UpperCAmelCase_ ): number += 1 while number < p_number_a: ans.append(UpperCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(UpperCAmelCase_ ): number += 1 # precondition assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ans[0] != p_number_a and ans[len(UpperCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def __lowerCamelCase ( UpperCAmelCase_ : Dict ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" a :Dict = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(UpperCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(UpperCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" a :int = get_divisors(UpperCAmelCase_ ) # precondition assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(UpperCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ): """simple docstring""" assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. a :Tuple = gcd(abs(UpperCAmelCase_ ) , abs(UpperCAmelCase_ ) ) # precondition assert ( isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def __lowerCamelCase ( UpperCAmelCase_ : Any ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" a :str = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def __lowerCamelCase ( UpperCAmelCase_ : int ): """simple docstring""" assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" a :List[str] = 0 a :List[str] = 1 a :int = 1 # this will be return for _ in range(n - 1 ): a :Tuple = ans ans += fiba a :Optional[Any] = tmp return ans
281
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case : List[str] = logging.get_logger(__name__) snake_case : Optional[Any] = { '''vocab_file''': '''vocab.txt''', '''merges_file''': '''bpe.codes''', } snake_case : str = { '''vocab_file''': { '''vinai/phobert-base''': '''https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt''', '''vinai/phobert-large''': '''https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt''', }, '''merges_file''': { '''vinai/phobert-base''': '''https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes''', '''vinai/phobert-large''': '''https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes''', }, } snake_case : List[Any] = { '''vinai/phobert-base''': 2_56, '''vinai/phobert-large''': 2_56, } def __lowerCamelCase ( UpperCAmelCase_ : List[str] ): """simple docstring""" a :Union[str, Any] = set() a :str = word[0] for char in word[1:]: pairs.add((prev_char, char) ) a :Optional[int] = char a :Optional[int] = set(UpperCAmelCase_ ) return pairs class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase="<s>" , _lowerCamelCase="</s>" , _lowerCamelCase="</s>" , _lowerCamelCase="<s>" , _lowerCamelCase="<unk>" , _lowerCamelCase="<pad>" , _lowerCamelCase="<mask>" , **_lowerCamelCase , ): super().__init__( bos_token=_lowerCamelCase , eos_token=_lowerCamelCase , unk_token=_lowerCamelCase , sep_token=_lowerCamelCase , cls_token=_lowerCamelCase , pad_token=_lowerCamelCase , mask_token=_lowerCamelCase , **_lowerCamelCase , ) a :Optional[Any] = vocab_file a :Optional[Any] = merges_file a :Any = {} a :Any = 0 a :int = 1 a :Union[str, Any] = 2 a :List[Any] = 3 self.add_from_file(_lowerCamelCase ) a :List[str] = {v: k for k, v in self.encoder.items()} with open(_lowerCamelCase , encoding='''utf-8''' ) as merges_handle: a :List[str] = merges_handle.read().split('''\n''' )[:-1] a :Any = [tuple(merge.split()[:-1] ) for merge in merges] a :str = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) a :str = {} def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a :Union[str, Any] = [self.cls_token_id] a :Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowerCamelCase , token_ids_a=_lowerCamelCase , already_has_special_tokens=_lowerCamelCase ) if token_ids_a is None: return [1] + ([0] * len(_lowerCamelCase )) + [1] return [1] + ([0] * len(_lowerCamelCase )) + [1, 1] + ([0] * len(_lowerCamelCase )) + [1] def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase = None ): a :Optional[int] = [self.sep_token_id] a :Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.encoder ) def SCREAMING_SNAKE_CASE__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): if token in self.cache: return self.cache[token] a :Optional[int] = tuple(_lowerCamelCase ) a :List[str] = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) a :Union[str, Any] = get_pairs(_lowerCamelCase ) if not pairs: return token while True: a :Optional[Any] = min(_lowerCamelCase , key=lambda _lowerCamelCase : self.bpe_ranks.get(_lowerCamelCase , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break a , a :Dict = bigram a :Union[str, Any] = [] a :int = 0 while i < len(_lowerCamelCase ): try: a :Optional[Any] = word.index(_lowerCamelCase , _lowerCamelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) a :Union[str, Any] = j if word[i] == first and i < len(_lowerCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 a :Union[str, Any] = tuple(_lowerCamelCase ) a :int = new_word if len(_lowerCamelCase ) == 1: break else: a :List[str] = get_pairs(_lowerCamelCase ) a :Union[str, Any] = '''@@ '''.join(_lowerCamelCase ) a :Dict = word[:-4] a :Any = word return word def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): a :Union[str, Any] = [] a :str = re.findall(R'''\S+\n?''' , _lowerCamelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCamelCase ).split(''' ''' ) ) ) return split_tokens def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return self.encoder.get(_lowerCamelCase , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return self.decoder.get(_lowerCamelCase , self.unk_token ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): a :Optional[int] = ''' '''.join(_lowerCamelCase ).replace('''@@ ''' , '''''' ).strip() return out_string def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase = None ): if not os.path.isdir(_lowerCamelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return a :Tuple = os.path.join( _lowerCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) a :Optional[int] = os.path.join( _lowerCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_lowerCamelCase ): copyfile(self.vocab_file , _lowerCamelCase ) if os.path.abspath(self.merges_file ) != os.path.abspath(_lowerCamelCase ): copyfile(self.merges_file , _lowerCamelCase ) return out_vocab_file, out_merge_file def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): if isinstance(_lowerCamelCase , _lowerCamelCase ): try: with open(_lowerCamelCase , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(_lowerCamelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return a :str = f.readlines() for lineTmp in lines: a :Tuple = lineTmp.strip() a :int = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) a :Tuple = line[:idx] a :Tuple = len(self.encoder )
281
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """TsinghuaAI/CPM-Generate""": """https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model""", } } class __snake_case ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , __lowerCamelCase , __lowerCamelCase=False , __lowerCamelCase=True , __lowerCamelCase=False , __lowerCamelCase="<s>" , __lowerCamelCase="</s>" , __lowerCamelCase="<unk>" , __lowerCamelCase="<sep>" , __lowerCamelCase="<pad>" , __lowerCamelCase="<cls>" , __lowerCamelCase="<mask>" , __lowerCamelCase=["<eop>", "<eod>"] , __lowerCamelCase = None , **__lowerCamelCase , ): '''simple docstring''' __A : Optional[int] = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token __A : Dict = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __A : Tuple = 3 __A : Union[str, Any] = do_lower_case __A : Dict = remove_space __A : List[str] = keep_accents __A : str = vocab_file __A : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( '''You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ''' '''See https://pypi.org/project/jieba/ for installation.''' ) __A : List[Any] = jieba __A : Optional[Any] = str.maketrans(''' \n''' , '''\u2582\u2583''' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def UpperCamelCase__( self ): '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__( self ): '''simple docstring''' __A : Any = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __A : int = self.__dict__.copy() __A : Union[str, Any] = None return state def __setstate__( self , __lowerCamelCase ): '''simple docstring''' __A : Union[str, Any] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __A : Union[str, Any] = {} __A : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' if self.remove_space: __A : List[str] = ' '.join(inputs.strip().split() ) else: __A : Optional[int] = inputs __A : List[str] = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __A : List[str] = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __A : int = ''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __A : List[Any] = outputs.lower() return outputs def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : List[Any] = self.preprocess_text(__UpperCAmelCase ) __A : Tuple = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __A : int = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __A : str = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __A : Dict = cur_pieces[1:] else: __A : Tuple = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : List[Any] = ''.join(__UpperCAmelCase ).replace(__UpperCAmelCase , ''' ''' ).strip() return out_string def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase = None ): '''simple docstring''' __A : List[str] = [self.sep_token_id] __A : List[Any] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase = None , __lowerCamelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] return ([0] * len(__UpperCAmelCase )) + [1, 1] def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase = None ): '''simple docstring''' __A : Any = [self.sep_token_id] __A : Union[str, Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __A : int = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __A : Dict = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,) def UpperCamelCase__( self , *__lowerCamelCase , **__lowerCamelCase ): '''simple docstring''' __A : List[str] = super()._decode(*__UpperCAmelCase , **__UpperCAmelCase ) __A : int = text.replace(''' ''' , '''''' ).replace('''\u2582''' , ''' ''' ).replace('''\u2583''' , '''\n''' ) return text
179
"""simple docstring""" import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel __A = logging.getLogger(__name__) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: """simple docstring""" if os.path.exists(_SCREAMING_SNAKE_CASE ): if os.path.exists(os.path.join(_SCREAMING_SNAKE_CASE , 'config.json' ) ) and os.path.isfile( os.path.join(_SCREAMING_SNAKE_CASE , 'config.json' ) ): os.remove(os.path.join(_SCREAMING_SNAKE_CASE , 'config.json' ) ) if os.path.exists(os.path.join(_SCREAMING_SNAKE_CASE , 'pytorch_model.bin' ) ) and os.path.isfile( os.path.join(_SCREAMING_SNAKE_CASE , 'pytorch_model.bin' ) ): os.remove(os.path.join(_SCREAMING_SNAKE_CASE , 'pytorch_model.bin' ) ) else: os.makedirs(_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) ->Optional[int]: """simple docstring""" lowerCAmelCase__ :Dict = 2 if unlogit: lowerCAmelCase__ :List[str] = torch.pow(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :str = p * torch.log(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[str] = 0 return -plogp.sum(dim=-1 ) def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" logger.info('lv, h >\t' + '\t'.join(F"{x + 1}" for x in range(len(_SCREAMING_SNAKE_CASE ) ) ) ) for row in range(len(_SCREAMING_SNAKE_CASE ) ): if tensor.dtype != torch.long: logger.info(F"layer {row + 1}:\t" + '\t'.join(F"{x:.5f}" for x in tensor[row].cpu().data ) ) else: logger.info(F"layer {row + 1}:\t" + '\t'.join(F"{x:d}" for x in tensor[row].cpu().data ) ) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False ) ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ :Dict = model.config.num_hidden_layers, model.config.num_attention_heads lowerCAmelCase__ :Any = torch.zeros(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).to(args.device ) lowerCAmelCase__ :Tuple = torch.zeros(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).to(args.device ) if head_mask is None: lowerCAmelCase__ :Optional[int] = torch.ones(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).to(args.device ) head_mask.requires_grad_(requires_grad=_SCREAMING_SNAKE_CASE ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: lowerCAmelCase__ :List[str] = None lowerCAmelCase__ :Any = 0.0 lowerCAmelCase__ :Any = 0.0 for step, inputs in enumerate(tqdm(_SCREAMING_SNAKE_CASE , desc='Iteration' , disable=args.local_rank not in [-1, 0] ) ): lowerCAmelCase__ :str = tuple(t.to(args.device ) for t in inputs ) ((lowerCAmelCase__) , ) :Dict = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) lowerCAmelCase__ :str = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE ) # (loss), lm_logits, presents, (all hidden_states), (attentions) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :str = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(_SCREAMING_SNAKE_CASE ): lowerCAmelCase__ :Optional[Any] = entropy(attn.detach() , _SCREAMING_SNAKE_CASE ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(_SCREAMING_SNAKE_CASE ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: lowerCAmelCase__ :Union[str, Any] = 2 lowerCAmelCase__ :Tuple = torch.pow(torch.pow(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-20 if not args.dont_normalize_global_importance: lowerCAmelCase__ :str = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('Attention entropies' ) print_ad_tensor(_SCREAMING_SNAKE_CASE ) if compute_importance: logger.info('Head importance scores' ) print_ad_tensor(_SCREAMING_SNAKE_CASE ) logger.info('Head ranked by importance scores' ) lowerCAmelCase__ :List[Any] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) lowerCAmelCase__ :List[Any] = torch.arange( head_importance.numel() , device=args.device ) lowerCAmelCase__ :int = head_ranks.view_as(_SCREAMING_SNAKE_CASE ) print_ad_tensor(_SCREAMING_SNAKE_CASE ) return attn_entropy, head_importance, total_loss def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = compute_heads_importance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , compute_entropy=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = 1 / loss # instead of downsteam score use the LM loss logger.info('Pruning: original score: %f, threshold: %f' , _SCREAMING_SNAKE_CASE , original_score * args.masking_threshold ) lowerCAmelCase__ :Optional[int] = torch.ones_like(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Dict = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) lowerCAmelCase__ :List[str] = original_score while current_score >= original_score * args.masking_threshold: lowerCAmelCase__ :List[str] = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads lowerCAmelCase__ :str = float('Inf' ) lowerCAmelCase__ :List[str] = head_importance.view(-1 ).sort()[1] if len(_SCREAMING_SNAKE_CASE ) <= num_to_mask: print('BREAK BY num_to_mask' ) break # mask heads lowerCAmelCase__ :int = current_heads_to_mask[:num_to_mask] logger.info('Heads to mask: %s' , str(current_heads_to_mask.tolist() ) ) lowerCAmelCase__ :Dict = new_head_mask.view(-1 ) lowerCAmelCase__ :Any = 0.0 lowerCAmelCase__ :Tuple = new_head_mask.view_as(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[int] = new_head_mask.clone().detach() print_ad_tensor(_SCREAMING_SNAKE_CASE ) # Compute metric and head importance again lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :Optional[Any] = compute_heads_importance( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , compute_entropy=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = 1 / loss logger.info( 'Masking: current score: %f, remaining heads %d (%.1f percents)' , _SCREAMING_SNAKE_CASE , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info('Final head mask' ) print_ad_tensor(_SCREAMING_SNAKE_CASE ) np.save(os.path.join(args.output_dir , 'head_mask.npy' ) , head_mask.detach().cpu().numpy() ) return head_mask def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: """simple docstring""" lowerCAmelCase__ :Union[str, Any] = datetime.now() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = compute_heads_importance( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , compute_entropy=_SCREAMING_SNAKE_CASE , compute_importance=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = 1 / loss lowerCAmelCase__ :Tuple = datetime.now() - before_time lowerCAmelCase__ :List[str] = sum(p.numel() for p in model.parameters() ) lowerCAmelCase__ :List[Any] = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(_SCREAMING_SNAKE_CASE ) ) } for k, v in heads_to_prune.items(): if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ :Union[str, Any] = [ v, ] assert sum(len(_SCREAMING_SNAKE_CASE ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = sum(p.numel() for p in model.parameters() ) lowerCAmelCase__ :int = datetime.now() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :Dict = compute_heads_importance( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , compute_entropy=_SCREAMING_SNAKE_CASE , compute_importance=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE , actually_pruned=_SCREAMING_SNAKE_CASE , ) lowerCAmelCase__ :int = 1 / loss lowerCAmelCase__ :Tuple = datetime.now() - before_time logger.info( 'Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , pruned_num_params / original_num_params * 100 , ) logger.info('Pruning: score with masking: %f score with pruning: %f' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) logger.info('Pruning: speed ratio (original timing / new timing): %f percents' , original_time / new_time * 100 ) save_model(_SCREAMING_SNAKE_CASE , args.output_dir ) def __A () ->Optional[Any]: """simple docstring""" lowerCAmelCase__ :List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--data_dir' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='The input data dir. Should contain the .tsv files (or other data files) for the task.' , ) parser.add_argument( '--model_name_or_path' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--output_dir' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , help='The output directory where the model predictions and checkpoints will be written.' , ) # Other parameters parser.add_argument( '--config_name' , default='' , type=_SCREAMING_SNAKE_CASE , help='Pretrained config name or path if not the same as model_name_or_path' , ) parser.add_argument( '--tokenizer_name' , default='' , type=_SCREAMING_SNAKE_CASE , help='Pretrained tokenizer name or path if not the same as model_name_or_path' , ) parser.add_argument( '--cache_dir' , default=_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , help='Where do you want to store the pre-trained models downloaded from s3' , ) parser.add_argument( '--data_subset' , type=_SCREAMING_SNAKE_CASE , default=-1 , help='If > 0: limit the data to a subset of data_subset instances.' ) parser.add_argument( '--overwrite_output_dir' , action='store_true' , help='Whether to overwrite data in output directory' ) parser.add_argument( '--overwrite_cache' , action='store_true' , help='Overwrite the cached training and evaluation sets' ) parser.add_argument( '--dont_normalize_importance_by_layer' , action='store_true' , help='Don\'t normalize importance score by layers' ) parser.add_argument( '--dont_normalize_global_importance' , action='store_true' , help='Don\'t normalize all importance scores between 0 and 1' , ) parser.add_argument( '--try_masking' , action='store_true' , help='Whether to try to mask head until a threshold of accuracy.' ) parser.add_argument( '--masking_threshold' , default=0.9 , type=_SCREAMING_SNAKE_CASE , help='masking threshold in term of metrics (stop masking when metric < threshold * original metric value).' , ) parser.add_argument( '--masking_amount' , default=0.1 , type=_SCREAMING_SNAKE_CASE , help='Amount to heads to masking at each masking step.' ) parser.add_argument('--metric_name' , default='acc' , type=_SCREAMING_SNAKE_CASE , help='Metric to use for head masking.' ) parser.add_argument( '--max_seq_length' , default=128 , type=_SCREAMING_SNAKE_CASE , help=( 'The maximum total input sequence length after WordPiece tokenization. \n' 'Sequences longer than this will be truncated, sequences shorter padded.' ) , ) parser.add_argument('--batch_size' , default=1 , type=_SCREAMING_SNAKE_CASE , help='Batch size.' ) parser.add_argument('--seed' , type=_SCREAMING_SNAKE_CASE , default=42 ) parser.add_argument('--local_rank' , type=_SCREAMING_SNAKE_CASE , default=-1 , help='local_rank for distributed training on gpus' ) parser.add_argument('--no_cuda' , action='store_true' , help='Whether not to use CUDA when available' ) parser.add_argument('--server_ip' , type=_SCREAMING_SNAKE_CASE , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=_SCREAMING_SNAKE_CASE , default='' , help='Can be used for distant debugging.' ) lowerCAmelCase__ :Any = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_SCREAMING_SNAKE_CASE ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: lowerCAmelCase__ :List[Any] = torch.device('cuda' if torch.cuda.is_available() and not args.no_cuda else 'cpu' ) lowerCAmelCase__ :Optional[int] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) lowerCAmelCase__ :Dict = torch.device('cuda' , args.local_rank ) lowerCAmelCase__ :Tuple = 1 torch.distributed.init_process_group(backend='nccl' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('device: {} n_gpu: {}, distributed: {}'.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) lowerCAmelCase__ :int = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: lowerCAmelCase__ :Optional[Any] = nn.parallel.DistributedDataParallel( _SCREAMING_SNAKE_CASE , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=_SCREAMING_SNAKE_CASE ) elif args.n_gpu > 1: lowerCAmelCase__ :Union[str, Any] = nn.DataParallel(_SCREAMING_SNAKE_CASE ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=_SCREAMING_SNAKE_CASE ) torch.save(_SCREAMING_SNAKE_CASE , os.path.join(args.output_dir , 'run_args.bin' ) ) logger.info('Training/evaluation parameters %s' , _SCREAMING_SNAKE_CASE ) # Prepare dataset lowerCAmelCase__ :Optional[int] = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) lowerCAmelCase__ :Union[str, Any] = (torch.from_numpy(_SCREAMING_SNAKE_CASE ),) lowerCAmelCase__ :Optional[int] = TensorDataset(*_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = RandomSampler(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Dict = DataLoader(_SCREAMING_SNAKE_CASE , sampler=_SCREAMING_SNAKE_CASE , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: lowerCAmelCase__ :Optional[Any] = mask_heads(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) prune_heads(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
293
0
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Input value must be an 'int' type" ) __SCREAMING_SNAKE_CASE = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
195
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a__ : Dict = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a__ : Any = TaTokenizerFast a__ : Tuple = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Union[str, Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a__ : str = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
195
1
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowercase_ ( _lowerCamelCase : Tuple , _lowerCamelCase : List[str] , _lowerCamelCase : Optional[Any]): lowercase__ : str = 1.5 lowercase__ : Any = int(factor * num_class_images) lowercase__ : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowerCamelCase , aesthetic_weight=0.1) os.makedirs(f'''{class_data_dir}/images''' , exist_ok=_lowerCamelCase) if len(list(Path(f'''{class_data_dir}/images''').iterdir())) >= num_class_images: return while True: lowercase__ : Dict = client.query(text=_lowerCamelCase) if len(_lowerCamelCase) >= factor * num_class_images or num_images > 1E4: break else: lowercase__ : List[Any] = int(factor * num_images) lowercase__ : Any = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowerCamelCase , aesthetic_weight=0.1 , ) lowercase__ : List[str] = 0 lowercase__ : Dict = 0 lowercase__ : int = tqdm(desc="downloading real regularization images" , total=_lowerCamelCase) with open(f'''{class_data_dir}/caption.txt''' , "w") as fa, open(f'''{class_data_dir}/urls.txt''' , "w") as fa, open( f'''{class_data_dir}/images.txt''' , "w") as fa: while total < num_class_images: lowercase__ : List[str] = class_images[count] count += 1 try: lowercase__ : Union[str, Any] = requests.get(images["url"]) if img.status_code == 200: lowercase__ : List[str] = Image.open(BytesIO(img.content)) with open(f'''{class_data_dir}/images/{total}.jpg''' , "wb") as f: f.write(img.content) fa.write(images["caption"] + "\n") fa.write(images["url"] + "\n") fa.write(f'''{class_data_dir}/images/{total}.jpg''' + "\n") total += 1 pbar.update(1) else: continue except Exception: continue return def lowercase_ ( ): lowercase__ : Optional[int] = argparse.ArgumentParser("" , add_help=_lowerCamelCase) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowerCamelCase , type=_lowerCamelCase) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowerCamelCase , type=_lowerCamelCase) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowerCamelCase) return parser.parse_args() if __name__ == "__main__": UpperCamelCase = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
87
"""simple docstring""" def _snake_case ( UpperCAmelCase_ : Dict ): # noqa: E741 A__ = len(UpperCAmelCase_ ) A__ = 0 A__ = [0] * n A__ = [False] * n A__ = [False] * n def dfs(UpperCAmelCase_ : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any] ): if parent == root: out_edge_count += 1 A__ = True A__ = at for to in l[at]: if to == parent: pass elif not visited[to]: A__ = dfs(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) A__ = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: A__ = True # AP found via cycle if at == low[to]: A__ = True else: A__ = min(low[at] , UpperCAmelCase_ ) return out_edge_count for i in range(UpperCAmelCase_ ): if not visited[i]: A__ = 0 A__ = dfs(UpperCAmelCase_ , UpperCAmelCase_ , -1 , UpperCAmelCase_ ) A__ = out_edge_count > 1 for x in range(len(UpperCAmelCase_ ) ): if is_art[x] is True: print(UpperCAmelCase_ ) # Adjacency list of graph SCREAMING_SNAKE_CASE_ : Optional[int] = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
335
0
"""simple docstring""" import numpy as np def _snake_case ( lowercase__ ): return 1 / (1 + np.exp(-vector )) def _snake_case ( lowercase__ ): return vector * sigmoid(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
12
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[int] = hf_hub_download( repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Tuple = VideoClassificationPipeline(model=lowercase , image_processor=lowercase , top_k=2 ) _lowerCamelCase : List[str] = [ example_video_filepath, 'https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4', ] return video_classifier, examples def A_ ( self , lowercase , lowercase ): for example in examples: _lowerCamelCase : Tuple = video_classifier(lowercase ) self.assertEqual( lowercase , [ {'score': ANY(lowercase ), 'label': ANY(lowercase )}, {'score': ANY(lowercase ), 'label': ANY(lowercase )}, ] , ) @require_torch def A_ ( self ): _lowerCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-VideoMAEForVideoClassification' _lowerCamelCase : Tuple = VideoMAEFeatureExtractor( size={'shortest_edge': 10} , crop_size={'height': 10, 'width': 10} ) _lowerCamelCase : Dict = pipeline( 'video-classification' , model=lowercase , feature_extractor=lowercase , frame_sampling_rate=4 ) _lowerCamelCase : Any = hf_hub_download(repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Dict = video_classifier(lowercase , top_k=2 ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}] , ) _lowerCamelCase : str = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [ [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], ] , ) @require_tf def A_ ( self ): pass
12
1
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging a_ = logging.get_logger(__name__) class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =['input_features', 'attention_mask'] def __init__( self : List[str] , a : List[str]=80 , a : int=1_6000 , a : List[str]=0.0 , a : Optional[Any]=10 , a : int=25 , a : List[Any]="hamming_window" , a : Any=3_2768.0 , a : Optional[int]=0.97 , a : Optional[int]=1.0 , a : Any=True , a : Union[str, Any]=True , a : str=False , **a : List[str] , ) -> Dict: """simple docstring""" super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) SCREAMING_SNAKE_CASE : Tuple = feature_size SCREAMING_SNAKE_CASE : Tuple = sampling_rate SCREAMING_SNAKE_CASE : Tuple = padding_value SCREAMING_SNAKE_CASE : Optional[Any] = hop_length SCREAMING_SNAKE_CASE : List[str] = win_length SCREAMING_SNAKE_CASE : Optional[int] = frame_signal_scale SCREAMING_SNAKE_CASE : Dict = preemphasis_coeff SCREAMING_SNAKE_CASE : str = mel_floor SCREAMING_SNAKE_CASE : Optional[int] = normalize_means SCREAMING_SNAKE_CASE : Optional[int] = normalize_vars SCREAMING_SNAKE_CASE : Tuple = win_function SCREAMING_SNAKE_CASE : Optional[int] = return_attention_mask SCREAMING_SNAKE_CASE : List[str] = win_length * sampling_rate // 1000 SCREAMING_SNAKE_CASE : Any = hop_length * sampling_rate // 1000 SCREAMING_SNAKE_CASE : Optional[Any] = optimal_fft_length(self.sample_size ) SCREAMING_SNAKE_CASE : Optional[Any] = (self.n_fft // 2) + 1 def __UpperCamelCase ( self : int , a : np.array ) -> np.ndarray: """simple docstring""" if self.win_function == "hamming_window": SCREAMING_SNAKE_CASE : Optional[Any] = window_function(window_length=self.sample_size , name=self.win_function , periodic=a ) else: SCREAMING_SNAKE_CASE : int = window_function(window_length=self.sample_size , name=self.win_function ) SCREAMING_SNAKE_CASE : Any = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) SCREAMING_SNAKE_CASE : Optional[int] = spectrogram( one_waveform * self.frame_signal_scale , window=a , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=a , preemphasis=self.preemphasis_coeff , mel_filters=a , mel_floor=self.mel_floor , log_mel="log" , ) return msfc_features.T def __UpperCamelCase ( self : Any , a : Union[str, Any] , a : List[str] , a : int ) -> List[str]: """simple docstring""" if self.normalize_means: SCREAMING_SNAKE_CASE : Dict = x[:input_length].mean(axis=0 ) SCREAMING_SNAKE_CASE : Any = np.subtract(a , a ) if self.normalize_vars: SCREAMING_SNAKE_CASE : int = x[:input_length].std(axis=0 ) SCREAMING_SNAKE_CASE : Union[str, Any] = np.divide(a , a ) if input_length < x.shape[0]: SCREAMING_SNAKE_CASE : List[str] = padding_value # make sure array is in float32 SCREAMING_SNAKE_CASE : List[str] = x.astype(np.floataa ) return x def __UpperCamelCase ( self : int , a : List[np.ndarray] , a : Optional[np.ndarray] = None ) -> List[np.ndarray]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(a , a , self.padding_value ) for x, n in zip(a , a )] def __call__( self : Optional[int] , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Union[bool, str, PaddingStrategy] = False , a : Optional[int] = None , a : bool = False , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , **a : Union[str, Any] , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" F" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" F" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) SCREAMING_SNAKE_CASE : Optional[Any] = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) SCREAMING_SNAKE_CASE : Optional[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: SCREAMING_SNAKE_CASE : List[str] = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): SCREAMING_SNAKE_CASE : Any = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): SCREAMING_SNAKE_CASE : Dict = raw_speech.astype(np.floataa ) # always return batch if not is_batched: SCREAMING_SNAKE_CASE : Any = [raw_speech] # extract fbank features SCREAMING_SNAKE_CASE : Optional[int] = [self._extract_mfsc_features(a ) for one_waveform in raw_speech] # convert into correct format for padding SCREAMING_SNAKE_CASE : List[Any] = BatchFeature({"input_features": features} ) SCREAMING_SNAKE_CASE : str = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format SCREAMING_SNAKE_CASE : Optional[int] = padded_inputs.get("input_features" ) if isinstance(input_features[0] , a ): SCREAMING_SNAKE_CASE : Tuple = [np.asarray(a , dtype=np.floataa ) for feature in input_features] SCREAMING_SNAKE_CASE : Any = padded_inputs.get("attention_mask" ) if attention_mask is not None: SCREAMING_SNAKE_CASE : List[str] = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: SCREAMING_SNAKE_CASE : str = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) SCREAMING_SNAKE_CASE : List[str] = self.normalize( padded_inputs["input_features"] , attention_mask=a ) if return_tensors is not None: SCREAMING_SNAKE_CASE : Optional[Any] = padded_inputs.convert_to_tensors(a ) return padded_inputs
76
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): # initialize config if "resnet-50" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-50") elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-101") else: raise ValueError("Model name should include either resnet50 or resnet101") SCREAMING_SNAKE_CASE : str = DetrConfig(use_timm_backbone=_a , backbone_config=_a) # set label attributes SCREAMING_SNAKE_CASE : List[str] = "panoptic" in model_name if is_panoptic: SCREAMING_SNAKE_CASE : Union[str, Any] = 250 else: SCREAMING_SNAKE_CASE : Union[str, Any] = 91 SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : Union[str, Any] = "coco-detection-id2label.json" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(_a , _a , repo_type="dataset") , "r")) SCREAMING_SNAKE_CASE : int = {int(_a): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCamelCase__ ( _a): # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE : Union[str, Any] = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight")) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight")) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias")) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean")) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var")) # stages for stage_idx in range(len(config.backbone_config.depths)): for layer_idx in range(config.backbone_config.depths[stage_idx]): # shortcut if layer_idx == 0: rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var", )) # 3 convs for i in range(3): rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var", )) # fmt: on for i in range(config.encoder_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", )) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", )) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ]) return rename_keys def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : str = state_dict.pop(_a) SCREAMING_SNAKE_CASE : int = val def lowerCamelCase__ ( _a , _a=False): SCREAMING_SNAKE_CASE : Optional[Any] = "" if is_panoptic: SCREAMING_SNAKE_CASE : Optional[int] = "detr." # first: transformer encoder for i in range(6): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : int = in_proj_bias[:256] SCREAMING_SNAKE_CASE : Tuple = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : str = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : str = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[:256] SCREAMING_SNAKE_CASE : List[Any] = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : Any = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight") SCREAMING_SNAKE_CASE : int = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias") # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE : Tuple = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[-256:] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Union[str, Any] = Image.open(requests.get(_a , stream=_a).raw) return im @torch.no_grad() def lowerCamelCase__ ( _a , _a=None , _a=False): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Optional[int] = get_detr_config(_a) # load original model from torch hub SCREAMING_SNAKE_CASE : Union[str, Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f"Converting model {model_name}...") SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=_a).eval() SCREAMING_SNAKE_CASE : Tuple = detr.state_dict() # rename keys for src, dest in create_rename_keys(_a): if is_panoptic: SCREAMING_SNAKE_CASE : List[str] = "detr." + src rename_key(_a , _a , _a) # query, key and value matrices need special treatment read_in_q_k_v(_a , is_panoptic=_a) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE : List[Any] = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr") and not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor") ): SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Optional[int] = val elif key.startswith("bbox_attention") or key.startswith("mask_head"): continue else: SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : List[Any] = val else: if not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor"): SCREAMING_SNAKE_CASE : Any = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Any = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE : int = DetrForSegmentation(_a) if is_panoptic else DetrForObjectDetection(_a) model.load_state_dict(_a) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE : int = "coco_panoptic" if is_panoptic else "coco_detection" SCREAMING_SNAKE_CASE : Optional[int] = DetrImageProcessor(format=_a) SCREAMING_SNAKE_CASE : List[str] = processor(images=prepare_img() , return_tensors="pt") SCREAMING_SNAKE_CASE : Any = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = detr(_a) SCREAMING_SNAKE_CASE : Any = model(_a) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4) print("Looks ok!") if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}...") Path(_a).mkdir(exist_ok=_a) model.save_pretrained(_a) processor.save_pretrained(_a) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub...") model.push_to_hub(f"nielsr/{model_name}") processor.push_to_hub(f"nielsr/{model_name}") if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') a_ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
76
1
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _lowerCAmelCase ( UpperCAmelCase : Dict ): '''simple docstring''' UpperCamelCase__ : int ={} UpperCamelCase__ : Optional[Any] =tokenizer(example['''content'''] , truncation=UpperCAmelCase )['''input_ids'''] UpperCamelCase__ : Any =len(example['''content'''] ) / len(output['''input_ids'''] ) return output _SCREAMING_SNAKE_CASE : Dict = HfArgumentParser(PretokenizationArguments) _SCREAMING_SNAKE_CASE : int = parser.parse_args() if args.num_workers is None: _SCREAMING_SNAKE_CASE : Optional[Any] = multiprocessing.cpu_count() _SCREAMING_SNAKE_CASE : str = AutoTokenizer.from_pretrained(args.tokenizer_dir) _SCREAMING_SNAKE_CASE : Optional[Any] = time.time() _SCREAMING_SNAKE_CASE : Union[str, Any] = load_dataset(args.dataset_name, split="""train""") print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') _SCREAMING_SNAKE_CASE : Tuple = time.time() _SCREAMING_SNAKE_CASE : Optional[int] = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') _SCREAMING_SNAKE_CASE : str = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
157
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : List[Any] = { """microsoft/unispeech-large-1500h-cv""": ( """https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json""" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class __a ( snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = 'unispeech' def __init__( self : List[Any] , lowercase_ : Tuple=32 , lowercase_ : int=768 , lowercase_ : List[Any]=12 , lowercase_ : Optional[int]=12 , lowercase_ : Union[str, Any]=3072 , lowercase_ : Any="gelu" , lowercase_ : List[Any]=0.1 , lowercase_ : str=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[str]=0.0 , lowercase_ : Tuple=0.0 , lowercase_ : Dict=0.1 , lowercase_ : List[str]=0.1 , lowercase_ : List[str]=0.0_2 , lowercase_ : int=1e-5 , lowercase_ : Dict="group" , lowercase_ : Optional[Any]="gelu" , lowercase_ : List[Any]=(512, 512, 512, 512, 512, 512, 512) , lowercase_ : Union[str, Any]=(5, 2, 2, 2, 2, 2, 2) , lowercase_ : Union[str, Any]=(10, 3, 3, 3, 3, 2, 2) , lowercase_ : Any=False , lowercase_ : Dict=128 , lowercase_ : List[str]=16 , lowercase_ : Any=False , lowercase_ : Optional[Any]=True , lowercase_ : List[str]=0.0_5 , lowercase_ : int=10 , lowercase_ : Optional[Any]=2 , lowercase_ : List[str]=0.0 , lowercase_ : List[Any]=10 , lowercase_ : Union[str, Any]=0 , lowercase_ : Dict=320 , lowercase_ : Optional[Any]=2 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : Dict=100 , lowercase_ : Optional[int]=256 , lowercase_ : Dict=256 , lowercase_ : Optional[int]=0.1 , lowercase_ : str="mean" , lowercase_ : Union[str, Any]=False , lowercase_ : Any=False , lowercase_ : Union[str, Any]=256 , lowercase_ : List[str]=80 , lowercase_ : Dict=0 , lowercase_ : int=1 , lowercase_ : Union[str, Any]=2 , lowercase_ : Dict=0.5 , **lowercase_ : str , ): super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) UpperCamelCase__ : Dict =hidden_size UpperCamelCase__ : Optional[int] =feat_extract_norm UpperCamelCase__ : Dict =feat_extract_activation UpperCamelCase__ : Union[str, Any] =list(lowercase_ ) UpperCamelCase__ : int =list(lowercase_ ) UpperCamelCase__ : Tuple =list(lowercase_ ) UpperCamelCase__ : List[str] =conv_bias UpperCamelCase__ : Any =num_conv_pos_embeddings UpperCamelCase__ : List[Any] =num_conv_pos_embedding_groups UpperCamelCase__ : Optional[int] =len(self.conv_dim ) UpperCamelCase__ : Union[str, Any] =num_hidden_layers UpperCamelCase__ : Optional[Any] =intermediate_size UpperCamelCase__ : Any =hidden_act UpperCamelCase__ : List[Any] =num_attention_heads UpperCamelCase__ : List[Any] =hidden_dropout UpperCamelCase__ : List[Any] =attention_dropout UpperCamelCase__ : Tuple =activation_dropout UpperCamelCase__ : Any =feat_proj_dropout UpperCamelCase__ : Tuple =final_dropout UpperCamelCase__ : Tuple =layerdrop UpperCamelCase__ : int =layer_norm_eps UpperCamelCase__ : Optional[int] =initializer_range UpperCamelCase__ : Any =num_ctc_classes UpperCamelCase__ : Optional[int] =vocab_size UpperCamelCase__ : int =do_stable_layer_norm UpperCamelCase__ : Union[str, Any] =use_weighted_layer_sum UpperCamelCase__ : Tuple =classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCamelCase__ : List[Any] =apply_spec_augment UpperCamelCase__ : List[Any] =mask_time_prob UpperCamelCase__ : Optional[int] =mask_time_length UpperCamelCase__ : Dict =mask_time_min_masks UpperCamelCase__ : str =mask_feature_prob UpperCamelCase__ : Union[str, Any] =mask_feature_length UpperCamelCase__ : int =mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCamelCase__ : Optional[Any] =num_codevectors_per_group UpperCamelCase__ : Dict =num_codevector_groups UpperCamelCase__ : int =contrastive_logits_temperature UpperCamelCase__ : Tuple =feat_quantizer_dropout UpperCamelCase__ : List[str] =num_negatives UpperCamelCase__ : Dict =codevector_dim UpperCamelCase__ : Any =proj_codevector_dim UpperCamelCase__ : List[Any] =diversity_loss_weight # ctc loss UpperCamelCase__ : Tuple =ctc_loss_reduction UpperCamelCase__ : List[str] =ctc_zero_infinity # pretraining loss UpperCamelCase__ : Optional[Any] =replace_prob @property def _lowerCAmelCase ( self : List[str] ): return functools.reduce(operator.mul , self.conv_stride , 1 )
157
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ :str = logging.get_logger(__name__) A_ :Optional[int] = {'''vocab_file''': '''spiece.model'''} A_ :str = { '''vocab_file''': { '''TsinghuaAI/CPM-Generate''': '''https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model''', } } class __A ( a ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<sep>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<cls>" , lowerCamelCase__="<mask>" , lowerCamelCase__=["<eop>", "<eod>"] , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[int] =AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token __UpperCamelCase : str ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase__ , remove_space=lowerCamelCase__ , keep_accents=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , additional_special_tokens=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) __UpperCamelCase : Optional[int] =3 __UpperCamelCase : Dict =do_lower_case __UpperCamelCase : Dict =remove_space __UpperCamelCase : Optional[Any] =keep_accents __UpperCamelCase : Tuple =vocab_file __UpperCamelCase : Tuple =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( 'You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ' 'See https://pypi.org/project/jieba/ for installation.' ) __UpperCamelCase : Tuple =jieba __UpperCamelCase : Any =str.maketrans(' \n' , '\u2582\u2583' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def __lowercase ( self ): """simple docstring""" return len(self.sp_model ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ={self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.__dict__.copy() __UpperCamelCase : Any =None return state def __setstate__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[Any] =d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __UpperCamelCase : List[Any] ={} __UpperCamelCase : Union[str, Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" if self.remove_space: __UpperCamelCase : List[Any] =' '.join(inputs.strip().split() ) else: __UpperCamelCase : Any =inputs __UpperCamelCase : Optional[Any] =outputs.replace('``' , '"' ).replace('\'\'' , '"' ) if not self.keep_accents: __UpperCamelCase : List[str] =unicodedata.normalize('NFKD' , lowerCamelCase__ ) __UpperCamelCase : List[Any] =''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase__ )] ) if self.do_lower_case: __UpperCamelCase : List[Any] =outputs.lower() return outputs def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =self.preprocess_text(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =[] for piece in pieces: if len(lowerCamelCase__ ) > 1 and piece[-1] == str(',' ) and piece[-2].isdigit(): __UpperCamelCase : Union[str, Any] =self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase__ , '' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __UpperCamelCase : Optional[int] =cur_pieces[1:] else: __UpperCamelCase : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(lowerCamelCase__ ) else: new_pieces.append(lowerCamelCase__ ) return new_pieces def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" return self.sp_model.PieceToId(lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" return self.sp_model.IdToPiece(lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[int] =''.join(lowerCamelCase__ ).replace(lowerCamelCase__ , ' ' ).strip() return out_string def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : Optional[Any] =[self.sep_token_id] __UpperCamelCase : List[Any] =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is not None: return ([0] * len(lowerCamelCase__ )) + [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] return ([0] * len(lowerCamelCase__ )) + [1, 1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : List[str] =[self.sep_token_id] __UpperCamelCase : Any =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __UpperCamelCase : Optional[Any] =os.path.join( lowerCamelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , 'wb' ) as fi: __UpperCamelCase : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,) def __lowercase ( self , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : int =super()._decode(*lowerCamelCase__ , **lowerCamelCase__ ) __UpperCamelCase : Optional[int] =text.replace(' ' , '' ).replace('\u2582' , ' ' ).replace('\u2583' , '\n' ) return text
71
import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __lowerCAmelCase : Optional[int] ='\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __lowerCAmelCase : Optional[Any] ='\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __lowerCAmelCase : Dict ='\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _UpperCamelCase ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Optional[Any] = simple_accuracy(lowercase__ , lowercase__ ) __SCREAMING_SNAKE_CASE : List[str] = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Optional[int] = np.array(lowercase__ ) __SCREAMING_SNAKE_CASE : str = np.array(lowercase__ ) __SCREAMING_SNAKE_CASE : str = en_sentvecs.shape[0] # mean centering __SCREAMING_SNAKE_CASE : Tuple = en_sentvecs - np.mean(lowercase__ , axis=0 ) __SCREAMING_SNAKE_CASE : Optional[int] = in_sentvecs - np.mean(lowercase__ , axis=0 ) __SCREAMING_SNAKE_CASE : str = cdist(lowercase__ , lowercase__ , '''cosine''' ) __SCREAMING_SNAKE_CASE : int = np.array(range(lowercase__ ) ) __SCREAMING_SNAKE_CASE : Optional[Any] = sim.argsort(axis=1 )[:, :10] __SCREAMING_SNAKE_CASE : str = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): '''simple docstring''' def __magic_name__( self :Tuple ) -> Tuple: if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), '''references''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if self.config_name != '''cvit-mkb-clsr''' else None , ) def __magic_name__( self :List[str] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple ) -> str: if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(lowerCAmelCase__ , lowerCAmelCase__ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(lowerCAmelCase__ , lowerCAmelCase__ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(lowerCAmelCase__ , lowerCAmelCase__ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' )
9
0
'''simple docstring''' import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures __lowerCamelCase = logging.get_logger(__name__) @dataclass class A__ : lowercase = field(metadata={"help": "The name of the task to train on: " + ", ".join(glue_processors.keys() )} ) lowercase = field( metadata={"help": "The input data dir. Should contain the .tsv files (or other data files) for the task."} ) lowercase = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowercase = field( default=_snake_case , metadata={"help": "Overwrite the cached training and evaluation sets"} ) def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = self.task_name.lower() class A__ ( _snake_case ): lowercase = "train" lowercase = "dev" lowercase = "test" class A__ ( _snake_case ): lowercase = 42 lowercase = 42 lowercase = 42 def __init__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = Split.train , UpperCamelCase__ = None , ) -> List[Any]: '''simple docstring''' warnings.warn( """This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets """ """library. You can have a look at this example script for pointers: """ """https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py""" , UpperCamelCase__ , ) A_ = args A_ = glue_processors[args.task_name]() A_ = glue_output_modes[args.task_name] if isinstance(UpperCamelCase__ , UpperCamelCase__ ): try: A_ = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) # Load data features from cache or dataset file A_ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}''' , ) A_ = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) A_ , A_ = label_list[2], label_list[1] A_ = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. A_ = cached_features_file + """.lock""" with FileLock(UpperCamelCase__ ): if os.path.exists(UpperCamelCase__ ) and not args.overwrite_cache: A_ = time.time() A_ = torch.load(UpperCamelCase__ ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) else: logger.info(f'''Creating features from dataset file at {args.data_dir}''' ) if mode == Split.dev: A_ = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: A_ = self.processor.get_test_examples(args.data_dir ) else: A_ = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: A_ = examples[:limit_length] A_ = glue_convert_examples_to_features( UpperCamelCase__ , UpperCamelCase__ , max_length=args.max_seq_length , label_list=UpperCamelCase__ , output_mode=self.output_mode , ) A_ = time.time() torch.save(self.features , UpperCamelCase__ ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self ) -> str: '''simple docstring''' return len(self.features ) def __getitem__( self , UpperCamelCase__ ) -> InputFeatures: '''simple docstring''' return self.features[i] def snake_case_ ( self ) -> Dict: '''simple docstring''' return self.label_list
369
'''simple docstring''' from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, ) -> list[float]: A_ , A_ = coefficient_matrix.shape A_ , A_ = constant_matrix.shape if rowsa != colsa: A_ = F'''Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}''' raise ValueError(UpperCAmelCase__ ) if colsa != 1: A_ = F'''Constant matrix must be nx1 but received {rowsa}x{colsa}''' raise ValueError(UpperCAmelCase__ ) if rowsa != rowsa: A_ = ( """Coefficient and constant matrices dimensions must be nxn and nx1 but """ F'''received {rowsa}x{colsa} and {rowsa}x{colsa}''' ) raise ValueError(UpperCAmelCase__ ) if len(UpperCAmelCase__ ) != rowsa: A_ = ( """Number of initial values must be equal to number of rows in coefficient """ F'''matrix but received {len(UpperCAmelCase__ )} and {rowsa}''' ) raise ValueError(UpperCAmelCase__ ) if iterations <= 0: raise ValueError("""Iterations must be at least 1""" ) A_ = np.concatenate( (coefficient_matrix, constant_matrix), axis=1 ) A_ , A_ = table.shape strictly_diagonally_dominant(UpperCAmelCase__ ) # Iterates the whole matrix for given number of times for _ in range(UpperCAmelCase__ ): A_ = [] for row in range(UpperCAmelCase__ ): A_ = 0 for col in range(UpperCAmelCase__ ): if col == row: A_ = table[row][col] elif col == cols - 1: A_ = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] A_ = (temp + val) / denom new_val.append(UpperCAmelCase__ ) A_ = new_val return [float(UpperCAmelCase__ ) for i in new_val] def UpperCAmelCase__ ( UpperCAmelCase__ ) -> bool: A_ , A_ = table.shape A_ = True for i in range(0, UpperCAmelCase__ ): A_ = 0 for j in range(0, cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("""Coefficient matrix is not strictly diagonally dominant""" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
101
0
# Algorithm for the pigeonhole sorting def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" _UpperCAmelCase = min(__lowerCAmelCase ) # min() finds the minimum value _UpperCAmelCase = max(__lowerCAmelCase ) # max() finds the maximum value _UpperCAmelCase = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size _UpperCAmelCase = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. _UpperCAmelCase = 0 for count in range(__lowerCAmelCase ): while holes[count] > 0: holes[count] -= 1 _UpperCAmelCase = count + min_val i += 1 def __A ( )-> Dict: """simple docstring""" _UpperCAmelCase = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(__lowerCAmelCase ) print('Sorted order is:' , ' '.join(__lowerCAmelCase ) ) if __name__ == "__main__": main()
39
"""simple docstring""" from __future__ import annotations from functools import lru_cache from math import ceil _A : Optional[Any] = 1_00 _A : Optional[int] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _A : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def __magic_name__ ( __snake_case : int ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} lowercase : set[int] = set() lowercase : int lowercase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def __magic_name__ ( __snake_case : int = 5000 ) -> int | None: for number_to_partition in range(1 , __snake_case ): if len(partition(__snake_case ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F"{solution() = }")
202
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) _a = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
100
"""simple docstring""" def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) _UpperCamelCase = str(bin(__snake_case ) ) binary_number += "0" * shift_amount return binary_number def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) _UpperCamelCase = str(bin(__snake_case ) )[2:] if shift_amount >= len(__snake_case ): return "0b0" _UpperCamelCase = binary_number[: len(__snake_case ) - shift_amount] return "0b" + shifted_binary_number def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number >= 0: # Get binary representation of positive number _UpperCamelCase = '''0''' + str(bin(__snake_case ) ).strip('''-''' )[2:] else: # Get binary (2's complement) representation of negative number _UpperCamelCase = len(bin(__snake_case )[3:] ) # Find 2's complement of number _UpperCamelCase = bin(abs(__snake_case ) - (1 << binary_number_length) )[3:] _UpperCamelCase = ( '''1''' + '''0''' * (binary_number_length - len(__snake_case )) + binary_number ) if shift_amount >= len(__snake_case ): return "0b" + binary_number[0] * len(__snake_case ) return ( "0b" + binary_number[0] * shift_amount + binary_number[: len(__snake_case ) - shift_amount] ) if __name__ == "__main__": import doctest doctest.testmod()
100
1
'''simple docstring''' import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def A ( self : List[str] ): """simple docstring""" UpperCamelCase = FlaxMTaForConditionalGeneration.from_pretrained('google/mt5-small' ) UpperCamelCase = AutoTokenizer.from_pretrained('google/mt5-small' ) UpperCamelCase = tokenizer('Hello there' , return_tensors='np' ).input_ids UpperCamelCase = tokenizer('Hi I am' , return_tensors='np' ).input_ids UpperCamelCase = shift_tokens_right(UpperCamelCase__ , model.config.pad_token_id , model.config.decoder_start_token_id ) UpperCamelCase = model(UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ ).logits UpperCamelCase = optax.softmax_cross_entropy(UpperCamelCase__ , onehot(UpperCamelCase__ , logits.shape[-1] ) ).mean() UpperCamelCase = -(labels.shape[-1] * loss.item()) UpperCamelCase = -8_4.9_1_2_7 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
28
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={ "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[str] ): for attribute in key.split('.' ): lowercase_ : Tuple = getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if weight_type is not None: lowercase_ : List[Any] = getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).shape else: lowercase_ : str = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": lowercase_ : Optional[Any] = value elif weight_type == "weight_g": lowercase_ : Optional[Any] = value elif weight_type == "weight_v": lowercase_ : Optional[Any] = value elif weight_type == "bias": lowercase_ : Union[str, Any] = value else: lowercase_ : Any = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase_ : Optional[int] = [] lowercase_ : Optional[int] = fairseq_model.state_dict() lowercase_ : List[str] = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): lowercase_ : List[str] = False if "conv_layers" in name: load_conv_layer( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == 'group' , ) lowercase_ : str = True else: for key, mapped_key in MAPPING.items(): lowercase_ : int = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: lowercase_ : str = True if "*" in mapped_key: lowercase_ : int = name.split(__SCREAMING_SNAKE_CASE )[0].split('.' )[-2] lowercase_ : Optional[Any] = mapped_key.replace('*' , __SCREAMING_SNAKE_CASE ) if "weight_g" in name: lowercase_ : Any = 'weight_g' elif "weight_v" in name: lowercase_ : Tuple = 'weight_v' elif "weight" in name: lowercase_ : int = 'weight' elif "bias" in name: lowercase_ : List[Any] = 'bias' else: lowercase_ : Optional[Any] = None set_recursively(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(__SCREAMING_SNAKE_CASE ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase__( __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : str ): lowercase_ : Dict = full_name.split('conv_layers.' )[-1] lowercase_ : int = name.split('.' ) lowercase_ : Any = int(items[0] ) lowercase_ : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) lowercase_ : List[str] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) lowercase_ : List[Any] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) lowercase_ : Optional[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) lowercase_ : List[str] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__SCREAMING_SNAKE_CASE ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : int ): lowercase_ : str = SEWConfig() if is_finetuned: lowercase_ : List[Any] = model.wav_encoder.wav_model.cfg else: lowercase_ : Tuple = model.cfg lowercase_ : Any = fs_config.conv_bias lowercase_ : Optional[Any] = eval(fs_config.conv_feature_layers ) lowercase_ : int = [x[0] for x in conv_layers] lowercase_ : Any = [x[1] for x in conv_layers] lowercase_ : Optional[Any] = [x[2] for x in conv_layers] lowercase_ : Tuple = 'gelu' lowercase_ : str = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' lowercase_ : int = 0.0 lowercase_ : Any = fs_config.activation_fn.name lowercase_ : Tuple = fs_config.encoder_embed_dim lowercase_ : int = 0.02 lowercase_ : Union[str, Any] = fs_config.encoder_ffn_embed_dim lowercase_ : Tuple = 1E-5 lowercase_ : Union[str, Any] = fs_config.encoder_layerdrop lowercase_ : Tuple = fs_config.encoder_attention_heads lowercase_ : List[str] = fs_config.conv_pos_groups lowercase_ : Union[str, Any] = fs_config.conv_pos lowercase_ : str = len(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = fs_config.encoder_layers lowercase_ : str = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: lowercase_ : Dict = model.cfg lowercase_ : Dict = fs_config.final_dropout lowercase_ : Dict = fs_config.layerdrop lowercase_ : Optional[int] = fs_config.activation_dropout lowercase_ : Any = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 lowercase_ : List[Any] = fs_config.attention_dropout lowercase_ : Tuple = fs_config.dropout_input lowercase_ : List[Any] = fs_config.dropout lowercase_ : Any = fs_config.mask_channel_length lowercase_ : str = fs_config.mask_channel_prob lowercase_ : Optional[Any] = fs_config.mask_length lowercase_ : Tuple = fs_config.mask_prob lowercase_ : List[Any] = 'Wav2Vec2FeatureExtractor' lowercase_ : Union[str, Any] = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def lowercase__( __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : int=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Tuple=True ): if is_finetuned: lowercase_ , lowercase_ , lowercase_ : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: lowercase_ , lowercase_ , lowercase_ : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: lowercase_ : List[str] = SEWConfig.from_pretrained(__SCREAMING_SNAKE_CASE ) else: lowercase_ : Tuple = convert_config(model[0] , __SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = model[0].eval() lowercase_ : List[Any] = True if config.feat_extract_norm == 'layer' else False lowercase_ : Optional[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=__SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , ) if is_finetuned: if dict_path: lowercase_ : Dict = Dictionary.load(__SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowercase_ : str = target_dict.pad_index lowercase_ : Union[str, Any] = target_dict.bos_index lowercase_ : Tuple = target_dict.pad_index lowercase_ : List[Any] = target_dict.bos_index lowercase_ : Any = target_dict.eos_index lowercase_ : str = len(target_dict.symbols ) lowercase_ : Optional[Any] = os.path.join(__SCREAMING_SNAKE_CASE , 'vocab.json' ) if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__SCREAMING_SNAKE_CASE ) ) return os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) with open(__SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = WavaVecaCTCTokenizer( __SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__SCREAMING_SNAKE_CASE , ) lowercase_ : Tuple = WavaVecaProcessor(feature_extractor=__SCREAMING_SNAKE_CASE , tokenizer=__SCREAMING_SNAKE_CASE ) processor.save_pretrained(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = SEWForCTC(__SCREAMING_SNAKE_CASE ) else: lowercase_ : Any = SEWModel(__SCREAMING_SNAKE_CASE ) feature_extractor.save_pretrained(__SCREAMING_SNAKE_CASE ) recursively_load_weights(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) hf_model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE =argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) __SCREAMING_SNAKE_CASE =parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
213
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = """▁""" lowerCamelCase__ = {"""vocab_file""": """sentencepiece.bpe.model"""} lowerCamelCase__ = { """vocab_file""": { """facebook/mbart-large-en-ro""": ( """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model""" ), """facebook/mbart-large-cc25""": ( """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model""" ), } } lowerCamelCase__ = { """facebook/mbart-large-en-ro""": 1024, """facebook/mbart-large-cc25""": 1024, } # fmt: off lowerCamelCase__ = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN"""] class A__ ( snake_case_ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = ["input_ids", "attention_mask"] lowercase = [] lowercase = [] def __init__( self : Dict , a : Optional[int] , a : List[str]="<s>" , a : Optional[int]="</s>" , a : List[str]="</s>" , a : Any="<s>" , a : Optional[int]="<unk>" , a : Union[str, Any]="<pad>" , a : Any="<mask>" , a : str=None , a : Union[str, Any]=None , a : str=None , a : Optional[Dict[str, Any]] = None , a : Dict=None , **a : Union[str, Any] , ): '''simple docstring''' lowerCAmelCase__ : Tuple = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else mask_token lowerCAmelCase__ : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , mask_token=_A , tokenizer_file=_A , src_lang=_A , tgt_lang=_A , additional_special_tokens=_A , sp_model_kwargs=self.sp_model_kwargs , **_A , ) lowerCAmelCase__ : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_A ) ) lowerCAmelCase__ : int = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token lowerCAmelCase__ : Any = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab lowerCAmelCase__ : Optional[int] = 1 lowerCAmelCase__ : Optional[Any] = len(self.sp_model ) lowerCAmelCase__ : str = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(_A ) } lowerCAmelCase__ : List[Any] = {v: k for k, v in self.lang_code_to_id.items()} lowerCAmelCase__ : str = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) lowerCAmelCase__ : int = {v: k for k, v in self.fairseq_tokens_to_ids.items()} lowerCAmelCase__ : List[Any] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) lowerCAmelCase__ : List[str] = src_lang if src_lang is not None else 'en_XX' lowerCAmelCase__ : Optional[Any] = self.lang_code_to_id[self._src_lang] lowerCAmelCase__ : Optional[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = self.__dict__.copy() lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Optional[Any] , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase__ : List[Any] = {} lowerCAmelCase__ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' return self._src_lang @src_lang.setter def _lowerCamelCase ( self : Tuple , a : str ): '''simple docstring''' lowerCAmelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) lowerCAmelCase__ : Union[str, Any] = [1] * len(self.prefix_tokens ) lowerCAmelCase__ : Optional[Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(_A )) + suffix_ones return prefix_ones + ([0] * len(_A )) + ([0] * len(_A )) + suffix_ones def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowerCamelCase ( self : List[str] , a : Dict , a : str , a : Optional[str] , a : Optional[str] , **a : Optional[int] ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCAmelCase__ : List[str] = src_lang lowerCAmelCase__ : Optional[Any] = self(_A , add_special_tokens=_A , return_tensors=_A , **_A ) lowerCAmelCase__ : Tuple = self.convert_tokens_to_ids(_A ) lowerCAmelCase__ : Optional[Any] = tgt_lang_id return inputs def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = {self.convert_ids_to_tokens(_A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowerCamelCase ( self : Optional[int] , a : str ): '''simple docstring''' return self.sp_model.encode(_A , out_type=_A ) def _lowerCamelCase ( self : Optional[Any] , a : List[str] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowerCAmelCase__ : Dict = self.sp_model.PieceToId(_A ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _lowerCamelCase ( self : Union[str, Any] , a : List[str] ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _lowerCamelCase ( self : List[str] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = ''.join(_A ).replace(_A , ' ' ).strip() return out_string def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase__ : List[Any] = os.path.join( _A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _A ) elif not os.path.isfile(self.vocab_file ): with open(_A , 'wb' ) as fi: lowerCAmelCase__ : List[str] = self.sp_model.serialized_model_proto() fi.write(_A ) return (out_vocab_file,) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : str = "en_XX" , a : Optional[List[str]] = None , a : str = "ro_RO" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : int = src_lang lowerCAmelCase__ : Any = tgt_lang return super().prepare_seqaseq_batch(_A , _A , **_A ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCamelCase ( self : List[str] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.lang_code_to_id[src_lang] lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : int = [self.eos_token_id, self.cur_lang_code] def _lowerCamelCase ( self : str , a : str ): '''simple docstring''' lowerCAmelCase__ : int = self.lang_code_to_id[lang] lowerCAmelCase__ : Tuple = [] lowerCAmelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code]
350
import torch from torch import nn class A__ ( nn.Module ): def __init__( self : Optional[int] , a : Union[str, Any] , a : str , a : str , a : List[Any] , a : List[Any]=1 , a : Tuple=False ): '''simple docstring''' super().__init__() lowerCAmelCase__ : Dict = n_token lowerCAmelCase__ : Any = d_embed lowerCAmelCase__ : str = d_proj lowerCAmelCase__ : int = cutoffs + [n_token] lowerCAmelCase__ : Union[str, Any] = [0] + self.cutoffs lowerCAmelCase__ : str = div_val lowerCAmelCase__ : Tuple = self.cutoffs[0] lowerCAmelCase__ : Dict = len(self.cutoffs ) - 1 lowerCAmelCase__ : Any = self.shortlist_size + self.n_clusters if self.n_clusters > 0: lowerCAmelCase__ : int = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) lowerCAmelCase__ : Optional[Any] = nn.Parameter(torch.zeros(self.n_clusters ) ) lowerCAmelCase__ : Optional[int] = nn.ModuleList() lowerCAmelCase__ : Tuple = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(a , a ) ) ) else: self.out_projs.append(a ) self.out_layers.append(nn.Linear(a , a ) ) else: for i in range(len(self.cutoffs ) ): lowerCAmelCase__ , lowerCAmelCase__ : Any = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowerCAmelCase__ : Optional[Any] = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(a , a ) ) ) self.out_layers.append(nn.Linear(a , r_idx - l_idx ) ) lowerCAmelCase__ : Tuple = keep_order def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : int , a : List[str] , a : str ): '''simple docstring''' if proj is None: lowerCAmelCase__ : Tuple = nn.functional.linear(a , a , bias=a ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: lowerCAmelCase__ : int = nn.functional.linear(a , proj.t().contiguous() ) lowerCAmelCase__ : Tuple = nn.functional.linear(a , a , bias=a ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def _lowerCamelCase ( self : List[str] , a : List[Any] , a : Optional[int]=None , a : Tuple=False ): '''simple docstring''' if labels is not None: # Shift so that tokens < n predict n lowerCAmelCase__ : str = hidden[..., :-1, :].contiguous() lowerCAmelCase__ : Optional[Any] = labels[..., 1:].contiguous() lowerCAmelCase__ : List[Any] = hidden.view(-1 , hidden.size(-1 ) ) lowerCAmelCase__ : Tuple = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError('Input and labels should have the same size in the batch dimension.' ) else: lowerCAmelCase__ : Optional[Any] = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: lowerCAmelCase__ : Optional[Any] = self._compute_logit(a , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: lowerCAmelCase__ : str = labels != -100 lowerCAmelCase__ : int = torch.zeros_like(a , dtype=hidden.dtype , device=hidden.device ) lowerCAmelCase__ : List[str] = ( -nn.functional.log_softmax(a , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: lowerCAmelCase__ : Optional[Any] = nn.functional.log_softmax(a , dim=-1 ) else: # construct weights and biases lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowerCAmelCase__ : Any = self.out_layers[0].weight[l_idx:r_idx] lowerCAmelCase__ : Any = self.out_layers[0].bias[l_idx:r_idx] else: lowerCAmelCase__ : Optional[Any] = self.out_layers[i].weight lowerCAmelCase__ : Optional[int] = self.out_layers[i].bias if i == 0: lowerCAmelCase__ : Dict = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowerCAmelCase__ : Union[str, Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(a ) biases.append(a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = weights[0], biases[0], self.out_projs[0] lowerCAmelCase__ : List[Any] = self._compute_logit(a , a , a , a ) lowerCAmelCase__ : Union[str, Any] = nn.functional.log_softmax(a , dim=1 ) if labels is None: lowerCAmelCase__ : Tuple = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: lowerCAmelCase__ : Dict = torch.zeros_like(a , dtype=hidden.dtype , device=hidden.device ) lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : Union[str, Any] = [0] + self.cutoffs for i in range(len(a ) - 1 ): lowerCAmelCase__ , lowerCAmelCase__ : Tuple = cutoff_values[i], cutoff_values[i + 1] if labels is not None: lowerCAmelCase__ : Tuple = (labels >= l_idx) & (labels < r_idx) lowerCAmelCase__ : int = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue lowerCAmelCase__ : Tuple = labels.index_select(0 , a ) - l_idx lowerCAmelCase__ : Any = head_logprob.index_select(0 , a ) lowerCAmelCase__ : Optional[int] = hidden.index_select(0 , a ) else: lowerCAmelCase__ : Any = hidden if i == 0: if labels is not None: lowerCAmelCase__ : Union[str, Any] = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: lowerCAmelCase__ : List[str] = head_logprob[:, : self.cutoffs[0]] else: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = weights[i], biases[i], self.out_projs[i] lowerCAmelCase__ : Union[str, Any] = self._compute_logit(a , a , a , a ) lowerCAmelCase__ : Optional[int] = nn.functional.log_softmax(a , dim=1 ) lowerCAmelCase__ : List[Any] = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: lowerCAmelCase__ : List[str] = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: lowerCAmelCase__ : Tuple = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i lowerCAmelCase__ : Union[str, Any] = logprob_i if labels is not None: if (hasattr(self , 'keep_order' ) and self.keep_order) or keep_order: out.index_copy_(0 , a , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def _lowerCamelCase ( self : List[Any] , a : Any ): '''simple docstring''' if self.n_clusters == 0: lowerCAmelCase__ : Union[str, Any] = self._compute_logit(a , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(a , dim=-1 ) else: # construct weights and biases lowerCAmelCase__ , lowerCAmelCase__ : str = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowerCAmelCase__ : str = self.out_layers[0].weight[l_idx:r_idx] lowerCAmelCase__ : Dict = self.out_layers[0].bias[l_idx:r_idx] else: lowerCAmelCase__ : int = self.out_layers[i].weight lowerCAmelCase__ : int = self.out_layers[i].bias if i == 0: lowerCAmelCase__ : Optional[int] = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowerCAmelCase__ : Union[str, Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(a ) biases.append(a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : str = weights[0], biases[0], self.out_projs[0] lowerCAmelCase__ : Dict = self._compute_logit(a , a , a , a ) lowerCAmelCase__ : List[Any] = hidden.new_empty((head_logit.size(0 ), self.n_token) ) lowerCAmelCase__ : Optional[Any] = nn.functional.log_softmax(a , dim=1 ) lowerCAmelCase__ : List[Any] = [0] + self.cutoffs for i in range(len(a ) - 1 ): lowerCAmelCase__ , lowerCAmelCase__ : str = cutoff_values[i], cutoff_values[i + 1] if i == 0: lowerCAmelCase__ : Union[str, Any] = head_logprob[:, : self.cutoffs[0]] else: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = weights[i], biases[i], self.out_projs[i] lowerCAmelCase__ : Dict = self._compute_logit(a , a , a , a ) lowerCAmelCase__ : List[str] = nn.functional.log_softmax(a , dim=1 ) lowerCAmelCase__ : Dict = head_logprob[:, -i] + tail_logprob_i lowerCAmelCase__ : List[str] = logprob_i return out
307
0
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings @maybe_allow_in_graph class UpperCAmelCase_ ( nn.Module ): def __init__( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int=0.0 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : str = "geglu" , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : str = "layer_norm" , UpperCAmelCase__ : bool = False , ) -> List[Any]: super().__init__() lowerCAmelCase = only_cross_attention lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == 'ada_norm_zero' lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == 'ada_norm' if norm_type in ("ada_norm", "ada_norm_zero") and num_embeds_ada_norm is None: raise ValueError( F'''`norm_type` is set to {norm_type}, but `num_embeds_ada_norm` is not defined. Please make sure to''' F''' define `num_embeds_ada_norm` if setting `norm_type` to {norm_type}.''' ) # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn if self.use_ada_layer_norm: lowerCAmelCase = AdaLayerNorm(UpperCAmelCase__ , UpperCAmelCase__ ) elif self.use_ada_layer_norm_zero: lowerCAmelCase = AdaLayerNormZero(UpperCAmelCase__ , UpperCAmelCase__ ) else: lowerCAmelCase = nn.LayerNorm(UpperCAmelCase__ , elementwise_affine=UpperCAmelCase__ ) lowerCAmelCase = Attention( query_dim=UpperCAmelCase__ , heads=UpperCAmelCase__ , dim_head=UpperCAmelCase__ , dropout=UpperCAmelCase__ , bias=UpperCAmelCase__ , cross_attention_dim=cross_attention_dim if only_cross_attention else None , upcast_attention=UpperCAmelCase__ , ) # 2. Cross-Attn if cross_attention_dim is not None or double_self_attention: # We currently only use AdaLayerNormZero for self attention where there will only be one attention block. # I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during # the second cross attention block. lowerCAmelCase = ( AdaLayerNorm(UpperCAmelCase__ , UpperCAmelCase__ ) if self.use_ada_layer_norm else nn.LayerNorm(UpperCAmelCase__ , elementwise_affine=UpperCAmelCase__ ) ) lowerCAmelCase = Attention( query_dim=UpperCAmelCase__ , cross_attention_dim=cross_attention_dim if not double_self_attention else None , heads=UpperCAmelCase__ , dim_head=UpperCAmelCase__ , dropout=UpperCAmelCase__ , bias=UpperCAmelCase__ , upcast_attention=UpperCAmelCase__ , ) # is self-attn if encoder_hidden_states is none else: lowerCAmelCase = None lowerCAmelCase = None # 3. Feed-forward lowerCAmelCase = nn.LayerNorm(UpperCAmelCase__ , elementwise_affine=UpperCAmelCase__ ) lowerCAmelCase = FeedForward(UpperCAmelCase__ , dropout=UpperCAmelCase__ , activation_fn=UpperCAmelCase__ , final_dropout=UpperCAmelCase__ ) # let chunk size default to None lowerCAmelCase = None lowerCAmelCase = 0 def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int ) -> Optional[int]: # Sets chunk feed-forward lowerCAmelCase = chunk_size lowerCAmelCase = dim def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[torch.LongTensor] = None , UpperCAmelCase__ : Dict[str, Any] = None , UpperCAmelCase__ : Optional[torch.LongTensor] = None , ) -> Any: # Notice that normalization is always applied before the real computation in the following blocks. # 1. Self-Attention if self.use_ada_layer_norm: lowerCAmelCase = self.norma(UpperCAmelCase__ , UpperCAmelCase__ ) elif self.use_ada_layer_norm_zero: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self.norma( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , hidden_dtype=hidden_states.dtype ) else: lowerCAmelCase = self.norma(UpperCAmelCase__ ) lowerCAmelCase = cross_attention_kwargs if cross_attention_kwargs is not None else {} lowerCAmelCase = self.attna( UpperCAmelCase__ , encoder_hidden_states=encoder_hidden_states if self.only_cross_attention else None , attention_mask=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.use_ada_layer_norm_zero: lowerCAmelCase = gate_msa.unsqueeze(1 ) * attn_output lowerCAmelCase = attn_output + hidden_states # 2. Cross-Attention if self.attna is not None: lowerCAmelCase = ( self.norma(UpperCAmelCase__ , UpperCAmelCase__ ) if self.use_ada_layer_norm else self.norma(UpperCAmelCase__ ) ) lowerCAmelCase = self.attna( UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , **UpperCAmelCase__ , ) lowerCAmelCase = attn_output + hidden_states # 3. Feed-forward lowerCAmelCase = self.norma(UpperCAmelCase__ ) if self.use_ada_layer_norm_zero: lowerCAmelCase = norm_hidden_states * (1 + scale_mlp[:, None]) + shift_mlp[:, None] if self._chunk_size is not None: # "feed_forward_chunk_size" can be used to save memory if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0: raise ValueError( F'''`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`.''' ) lowerCAmelCase = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size lowerCAmelCase = torch.cat( [self.ff(UpperCAmelCase__ ) for hid_slice in norm_hidden_states.chunk(UpperCAmelCase__ , dim=self._chunk_dim )] , dim=self._chunk_dim , ) else: lowerCAmelCase = self.ff(UpperCAmelCase__ ) if self.use_ada_layer_norm_zero: lowerCAmelCase = gate_mlp.unsqueeze(1 ) * ff_output lowerCAmelCase = ff_output + hidden_states return hidden_states class UpperCAmelCase_ ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : int = 4 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : str = "geglu" , UpperCAmelCase__ : bool = False , ) -> str: super().__init__() lowerCAmelCase = int(dim * mult ) lowerCAmelCase = dim_out if dim_out is not None else dim if activation_fn == "gelu": lowerCAmelCase = GELU(UpperCAmelCase__ , UpperCAmelCase__ ) if activation_fn == "gelu-approximate": lowerCAmelCase = GELU(UpperCAmelCase__ , UpperCAmelCase__ , approximate='tanh' ) elif activation_fn == "geglu": lowerCAmelCase = GEGLU(UpperCAmelCase__ , UpperCAmelCase__ ) elif activation_fn == "geglu-approximate": lowerCAmelCase = ApproximateGELU(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = nn.ModuleList([] ) # project in self.net.append(UpperCAmelCase__ ) # project dropout self.net.append(nn.Dropout(UpperCAmelCase__ ) ) # project out self.net.append(nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) ) # FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout if final_dropout: self.net.append(nn.Dropout(UpperCAmelCase__ ) ) def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : Tuple ) -> str: for module in self.net: lowerCAmelCase = module(UpperCAmelCase__ ) return hidden_states class UpperCAmelCase_ ( nn.Module ): def __init__( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : str = "none" ) -> List[Any]: super().__init__() lowerCAmelCase = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = approximate def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : Optional[int] ) -> Any: if gate.device.type != "mps": return F.gelu(UpperCAmelCase__ , approximate=self.approximate ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) , approximate=self.approximate ).to(dtype=gate.dtype ) def __UpperCAmelCase ( self : Tuple , UpperCAmelCase__ : Tuple ) -> List[str]: lowerCAmelCase = self.proj(UpperCAmelCase__ ) lowerCAmelCase = self.gelu(UpperCAmelCase__ ) return hidden_states class UpperCAmelCase_ ( nn.Module ): def __init__( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> int: super().__init__() lowerCAmelCase = nn.Linear(UpperCAmelCase__ , dim_out * 2 ) def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: if gate.device.type != "mps": return F.gelu(UpperCAmelCase__ ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) ).to(dtype=gate.dtype ) def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : List[Any] ) -> Optional[int]: lowerCAmelCase , lowerCAmelCase = self.proj(UpperCAmelCase__ ).chunk(2 , dim=-1 ) return hidden_states * self.gelu(UpperCAmelCase__ ) class UpperCAmelCase_ ( nn.Module ): def __init__( self : Optional[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> Optional[Any]: super().__init__() lowerCAmelCase = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) def __UpperCAmelCase ( self : int , UpperCAmelCase__ : List[str] ) -> List[Any]: lowerCAmelCase = self.proj(UpperCAmelCase__ ) return x * torch.sigmoid(1.702 * x ) class UpperCAmelCase_ ( nn.Module ): def __init__( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any] ) -> Optional[int]: super().__init__() lowerCAmelCase = nn.Embedding(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = nn.SiLU() lowerCAmelCase = nn.Linear(UpperCAmelCase__ , embedding_dim * 2 ) lowerCAmelCase = nn.LayerNorm(UpperCAmelCase__ , elementwise_affine=UpperCAmelCase__ ) def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] ) -> Dict: lowerCAmelCase = self.linear(self.silu(self.emb(UpperCAmelCase__ ) ) ) lowerCAmelCase , lowerCAmelCase = torch.chunk(UpperCAmelCase__ , 2 ) lowerCAmelCase = self.norm(UpperCAmelCase__ ) * (1 + scale) + shift return x class UpperCAmelCase_ ( nn.Module ): def __init__( self : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int ) -> Dict: super().__init__() lowerCAmelCase = CombinedTimestepLabelEmbeddings(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = nn.SiLU() lowerCAmelCase = nn.Linear(UpperCAmelCase__ , 6 * embedding_dim , bias=UpperCAmelCase__ ) lowerCAmelCase = nn.LayerNorm(UpperCAmelCase__ , elementwise_affine=UpperCAmelCase__ , eps=1E-6 ) def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int]=None ) -> List[Any]: lowerCAmelCase = self.linear(self.silu(self.emb(UpperCAmelCase__ , UpperCAmelCase__ , hidden_dtype=UpperCAmelCase__ ) ) ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = emb.chunk(6 , dim=1 ) lowerCAmelCase = self.norm(UpperCAmelCase__ ) * (1 + scale_msa[:, None]) + shift_msa[:, None] return x, gate_msa, shift_mlp, scale_mlp, gate_mlp class UpperCAmelCase_ ( nn.Module ): def __init__( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : float = 1E-5 ) -> Optional[Any]: super().__init__() lowerCAmelCase = num_groups lowerCAmelCase = eps if act_fn is None: lowerCAmelCase = None else: lowerCAmelCase = get_activation(UpperCAmelCase__ ) lowerCAmelCase = nn.Linear(UpperCAmelCase__ , out_dim * 2 ) def __UpperCAmelCase ( self : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any ) -> Tuple: if self.act: lowerCAmelCase = self.act(UpperCAmelCase__ ) lowerCAmelCase = self.linear(UpperCAmelCase__ ) lowerCAmelCase = emb[:, :, None, None] lowerCAmelCase , lowerCAmelCase = emb.chunk(2 , dim=1 ) lowerCAmelCase = F.group_norm(UpperCAmelCase__ , self.num_groups , eps=self.eps ) lowerCAmelCase = x * (1 + scale) + shift return x
4
'''simple docstring''' import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class _lowerCAmelCase ( __UpperCAmelCase ): def __init__(self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = None , lowercase = None , **lowercase , ): super().__init__( lowercase , split=lowercase , features=lowercase , cache_dir=lowercase , keep_in_memory=lowercase , streaming=lowercase , num_proc=lowercase , **lowercase , ) A_ : Any = field A_ : Optional[int] = path_or_paths if isinstance(lowercase , lowercase ) else {self.split: path_or_paths} A_ : str = Json( cache_dir=lowercase , data_files=lowercase , features=lowercase , field=lowercase , **lowercase , ) def _a (self ): # Build iterable dataset if self.streaming: A_ : Optional[int] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: A_ : Optional[Any] = None A_ : Optional[Any] = None A_ : Tuple = None A_ : Optional[int] = None self.builder.download_and_prepare( download_config=lowercase , download_mode=lowercase , verification_mode=lowercase , base_path=lowercase , num_proc=self.num_proc , ) A_ : Union[str, Any] = self.builder.as_dataset( split=self.split , verification_mode=lowercase , in_memory=self.keep_in_memory ) return dataset class _lowerCAmelCase : def __init__(self , lowercase , lowercase , lowercase = None , lowercase = None , **lowercase , ): if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) A_ : Union[str, Any] = dataset A_ : Optional[int] = path_or_buf A_ : str = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE A_ : List[str] = num_proc A_ : Union[str, Any] = """utf-8""" A_ : Dict = to_json_kwargs def _a (self ): A_ : Optional[Any] = self.to_json_kwargs.pop("""path_or_buf""" , lowercase ) A_ : Tuple = self.to_json_kwargs.pop("""orient""" , """records""" ) A_ : Union[str, Any] = self.to_json_kwargs.pop("""lines""" , True if orient == """records""" else False ) A_ : str = self.to_json_kwargs.pop("""index""" , False if orient in ["""split""", """table"""] else True ) A_ : Dict = self.to_json_kwargs.pop("""compression""" , lowercase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F'`datasets` currently does not support {compression} compression' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , """wb""" , compression=lowercase ) as buffer: A_ : List[str] = self._write(file_obj=lowercase , orient=lowercase , lines=lowercase , index=lowercase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( F'The compression parameter is not supported when writing to a buffer, but compression={compression}' """ was passed. Please provide a local path instead.""" ) A_ : List[Any] = self._write( file_obj=self.path_or_buf , orient=lowercase , lines=lowercase , index=lowercase , **self.to_json_kwargs ) return written def _a (self , lowercase ): A_, A_, A_, A_, A_ : Dict = args A_ : Any = query_table( table=self.dataset.data , key=slice(lowercase , offset + self.batch_size ) , indices=self.dataset._indices , ) A_ : Union[str, Any] = batch.to_pandas().to_json( path_or_buf=lowercase , orient=lowercase , lines=lowercase , index=lowercase , **lowercase ) if not json_str.endswith("""\n""" ): json_str += "\n" return json_str.encode(self.encoding ) def _a (self , lowercase , lowercase , lowercase , lowercase , **lowercase , ): A_ : Optional[Any] = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): A_ : Any = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(lowercase ) else: A_, A_ : List[Any] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , lowercase , lowercase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): written += file_obj.write(lowercase ) return written
206
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase : Dict = { """configuration_instructblip""": [ """INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """InstructBlipConfig""", """InstructBlipQFormerConfig""", """InstructBlipVisionConfig""", ], """processing_instructblip""": ["""InstructBlipProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Union[str, Any] = [ """INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """InstructBlipQFormerModel""", """InstructBlipPreTrainedModel""", """InstructBlipForConditionalGeneration""", """InstructBlipVisionModel""", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys _lowercase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
91
'''simple docstring''' import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def lowerCamelCase__ ( A : str="" ): '''simple docstring''' UpperCAmelCase = tempfile.mkdtemp() return os.path.join(A , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class UpperCamelCase__( unittest.TestCase ): def a__( self : int )-> int: """simple docstring""" UpperCAmelCase = torch.rand(12 , dtype=torch.floataa ) - 0.5 UpperCAmelCase = AgentAudio(lowerCAmelCase ) UpperCAmelCase = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase , agent_type.to_raw() , atol=1E-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowerCAmelCase ) ) # Ensure that the file contains the same value as the original tensor UpperCAmelCase , UpperCAmelCase = sf.read(lowerCAmelCase ) self.assertTrue(torch.allclose(lowerCAmelCase , torch.tensor(lowerCAmelCase ) , atol=1E-4 ) ) def a__( self : Union[str, Any] )-> Optional[Any]: """simple docstring""" UpperCAmelCase = torch.rand(12 , dtype=torch.floataa ) - 0.5 UpperCAmelCase = get_new_path(suffix='''.wav''' ) sf.write(lowerCAmelCase , lowerCAmelCase , 16000 ) UpperCAmelCase = AgentAudio(lowerCAmelCase ) self.assertTrue(torch.allclose(lowerCAmelCase , agent_type.to_raw() , atol=1E-4 ) ) self.assertEqual(agent_type.to_string() , lowerCAmelCase ) @require_vision @require_torch class UpperCamelCase__( unittest.TestCase ): def a__( self : List[Any] )-> Any: """simple docstring""" UpperCAmelCase = torch.randint(0 , 256 , (64, 64, 3) ) UpperCAmelCase = AgentImage(lowerCAmelCase ) UpperCAmelCase = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase , agent_type._tensor , atol=1E-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase ) ) def a__( self : List[Any] )-> List[Any]: """simple docstring""" UpperCAmelCase = Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' UpperCAmelCase = Image.open(lowerCAmelCase ) UpperCAmelCase = AgentImage(lowerCAmelCase ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase ) ) def a__( self : Optional[Any] )-> List[str]: """simple docstring""" UpperCAmelCase = Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' UpperCAmelCase = Image.open(lowerCAmelCase ) UpperCAmelCase = AgentImage(lowerCAmelCase ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase ) ) class UpperCamelCase__( unittest.TestCase ): def a__( self : int )-> Any: """simple docstring""" UpperCAmelCase = '''Hey!''' UpperCAmelCase = AgentText(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , agent_type.to_string() ) self.assertEqual(lowerCAmelCase , agent_type.to_raw() ) self.assertEqual(lowerCAmelCase , lowerCAmelCase )
91
1
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str ) -> List[str]: """simple docstring""" __lowerCamelCase = 1.5 __lowerCamelCase = int(factor * num_class_images ) __lowerCamelCase = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=UpperCamelCase__ , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=UpperCamelCase__ ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: __lowerCamelCase = client.query(text=UpperCamelCase__ ) if len(UpperCamelCase__ ) >= factor * num_class_images or num_images > 1E4: break else: __lowerCamelCase = int(factor * num_images ) __lowerCamelCase = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=UpperCamelCase__ , aesthetic_weight=0.1 , ) __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(desc='downloading real regularization images' , total=UpperCamelCase__ ) with open(F"""{class_data_dir}/caption.txt""" , 'w' ) as fa, open(F"""{class_data_dir}/urls.txt""" , 'w' ) as fa, open( F"""{class_data_dir}/images.txt""" , 'w' ) as fa: while total < num_class_images: __lowerCamelCase = class_images[count] count += 1 try: __lowerCamelCase = requests.get(images['url'] ) if img.status_code == 200: __lowerCamelCase = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCamelCase_ ( ) -> Any: """simple docstring""" __lowerCamelCase = argparse.ArgumentParser('' , add_help=UpperCamelCase__ ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=UpperCamelCase__ , type=UpperCamelCase__ ) parser.add_argument('--class_data_dir' , help='path to save images' , required=UpperCamelCase__ , type=UpperCamelCase__ ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=UpperCamelCase__ ) return parser.parse_args() if __name__ == "__main__": __A = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
90
"""simple docstring""" import os def _snake_case ( ) -> Dict: with open(os.path.dirname(lowerCamelCase__ ) + "/p022_names.txt" ) as file: lowerCamelCase_ : str =str(file.readlines()[0] ) lowerCamelCase_ : Union[str, Any] =names.replace("\"" , "" ).split("," ) names.sort() lowerCamelCase_ : str =0 lowerCamelCase_ : Optional[int] =0 for i, name in enumerate(lowerCamelCase__ ): for letter in name: name_score += ord(lowerCamelCase__ ) - 64 total_score += (i + 1) * name_score lowerCamelCase_ : List[Any] =0 return total_score if __name__ == "__main__": print(solution())
144
0
from math import sqrt def lowerCamelCase__ ( a ) -> bool: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" _A: Any = True # 0 and 1 are none primes. if number <= 1: _A: Optional[int] = False for divisor in range(2 , int(round(sqrt(__UpperCAmelCase ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _A: Optional[Any] = False break # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'status' must been from type bool" return status def lowerCamelCase__ ( a ) -> Any: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _A: int = list(range(2 , n + 1 ) ) _A: int = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__UpperCAmelCase ) ): for j in range(i + 1 , len(__UpperCAmelCase ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _A: int = 0 # filters actual prime numbers. _A: Optional[Any] = [x for x in begin_list if x != 0] # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase__ ( a ) -> Dict: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n > 2), "'N' must been an int and > 2" _A: Optional[int] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(__UpperCAmelCase ): ans.append(__UpperCAmelCase ) # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase__ ( a ) -> Union[str, Any]: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and number >= 0, "'number' must been an int and >= 0" _A: List[Any] = [] # this list will be returns of the function. # potential prime number factors. _A: Union[str, Any] = 2 _A: Any = number if number == 0 or number == 1: ans.append(__UpperCAmelCase ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__UpperCAmelCase ): while quotient != 1: if is_prime(__UpperCAmelCase ) and (quotient % factor == 0): ans.append(__UpperCAmelCase ) quotient /= factor else: factor += 1 else: ans.append(__UpperCAmelCase ) # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase__ ( a ) -> Union[str, Any]: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" _A: Optional[int] = 0 # prime factorization of 'number' _A: int = prime_factorization(__UpperCAmelCase ) _A: int = max(__UpperCAmelCase ) # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'ans' must been from type int" return ans def lowerCamelCase__ ( a ) -> Tuple: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" _A: Tuple = 0 # prime factorization of 'number' _A: Dict = prime_factorization(__UpperCAmelCase ) _A: Optional[int] = min(__UpperCAmelCase ) # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'ans' must been from type int" return ans def lowerCamelCase__ ( a ) -> Dict: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'number' must been an int" assert isinstance(number % 2 == 0 , __UpperCAmelCase ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( a ) -> int: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ), "'number' must been an int" assert isinstance(number % 2 != 0 , __UpperCAmelCase ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( a ) -> Dict: assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (number > 2) and is_even(__UpperCAmelCase ) ), "'number' must been an int, even and > 2" _A: Any = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _A: Tuple = get_prime_numbers(__UpperCAmelCase ) _A: int = len(__UpperCAmelCase ) # run variable for while-loops. _A: List[Any] = 0 _A: Tuple = None # exit variable. for break up the loops _A: int = True while i < len_pn and loop: _A: Tuple = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _A: Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (len(__UpperCAmelCase ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( a , a ) -> Union[str, Any]: assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _A: Tuple = 0 while numbera != 0: _A: List[str] = numbera % numbera _A: Any = numbera _A: Tuple = rest # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( a , a ) -> Optional[int]: assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _A: Dict = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _A: List[Any] = prime_factorization(__UpperCAmelCase ) _A: Any = prime_factorization(__UpperCAmelCase ) elif numbera == 1 or numbera == 1: _A: Tuple = [] _A: Tuple = [] _A: List[str] = max(__UpperCAmelCase , __UpperCAmelCase ) _A: Optional[Any] = 0 _A: int = 0 _A: List[Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _A: str = prime_fac_a.count(__UpperCAmelCase ) _A: Optional[Any] = prime_fac_a.count(__UpperCAmelCase ) for _ in range(max(__UpperCAmelCase , __UpperCAmelCase ) ): ans *= n else: _A: List[str] = prime_fac_a.count(__UpperCAmelCase ) for _ in range(__UpperCAmelCase ): ans *= n done.append(__UpperCAmelCase ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _A: Tuple = prime_fac_a.count(__UpperCAmelCase ) for _ in range(__UpperCAmelCase ): ans *= n done.append(__UpperCAmelCase ) # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( a ) -> int: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n >= 0), "'number' must been a positive int" _A: int = 0 _A: Dict = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__UpperCAmelCase ): ans += 1 # precondition assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and is_prime( __UpperCAmelCase ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( a , a ) -> str: assert ( is_prime(__UpperCAmelCase ) and is_prime(__UpperCAmelCase ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _A: List[Any] = p_number_a + 1 # jump to the next number _A: Tuple = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__UpperCAmelCase ): number += 1 while number < p_number_a: ans.append(__UpperCAmelCase ) number += 1 # fetch the next prime number. while not is_prime(__UpperCAmelCase ): number += 1 # precondition assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ans[0] != p_number_a and ans[len(__UpperCAmelCase ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( a ) -> str: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n >= 1), "'n' must been int and >= 1" _A: List[str] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(__UpperCAmelCase ) # precondition assert ans[0] == 1 and ans[len(__UpperCAmelCase ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( a ) -> Dict: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and ( number > 1 ), "'number' must been an int and >= 1" _A: Tuple = get_divisors(__UpperCAmelCase ) # precondition assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (divisors[0] == 1) and (divisors[len(__UpperCAmelCase ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( a , a ) -> Optional[int]: assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _A: List[Any] = gcd(abs(__UpperCAmelCase ) , abs(__UpperCAmelCase ) ) # precondition assert ( isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( a ) -> Optional[int]: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n >= 0), "'n' must been a int and >= 0" _A: Optional[Any] = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCamelCase__ ( a ) -> Tuple: assert isinstance(__UpperCAmelCase , __UpperCAmelCase ) and (n >= 0), "'n' must been an int and >= 0" _A: Dict = 0 _A: int = 1 _A: int = 1 # this will be return for _ in range(n - 1 ): _A: int = ans ans += fiba _A: Tuple = tmp return ans
356
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params UpperCAmelCase__ : Any = getLogger(__name__) UpperCAmelCase__ : Optional[Any] = 'cuda' if torch.cuda.is_available() else 'cpu' def lowerCamelCase__ ( a , a , a , a = 8 , a = DEFAULT_DEVICE , a=False , a="summarization" , a=None , **a , ) -> Dict: _A: str = Path(a ).open('''w''' , encoding='''utf-8''' ) _A: Optional[Any] = str(a ) _A: Union[str, Any] = AutoModelForSeqaSeqLM.from_pretrained(a ).to(a ) if fpaa: _A: Any = model.half() _A: Optional[int] = AutoTokenizer.from_pretrained(a ) logger.info(f"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. _A: Any = time.time() # update config with task specific params use_task_specific_params(a , a ) if prefix is None: _A: int = prefix or getattr(model.config , '''prefix''' , '''''' ) or '''''' for examples_chunk in tqdm(list(chunks(a , a ) ) ): _A: int = [prefix + text for text in examples_chunk] _A: str = tokenizer(a , return_tensors='''pt''' , truncation=a , padding='''longest''' ).to(a ) _A: str = model.generate( input_ids=batch.input_ids , attention_mask=batch.attention_mask , **a , ) _A: str = tokenizer.batch_decode(a , skip_special_tokens=a , clean_up_tokenization_spaces=a ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() _A: Optional[int] = int(time.time() - start_time ) # seconds _A: Union[str, Any] = len(a ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs , 4 )} def lowerCamelCase__ ( ) -> Tuple: return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def lowerCamelCase__ ( a=True ) -> Optional[Any]: _A: str = argparse.ArgumentParser() parser.add_argument('''model_name''' , type=a , help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' , type=a , help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' , type=a , help='''where to save summaries''' ) parser.add_argument('''--reference_path''' , type=a , required=a , help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' , type=a , required=a , default='''metrics.json''' , help='''where to save metrics''' ) parser.add_argument('''--device''' , type=a , required=a , default=a , help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' , type=a , required=a , default=a , help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' , type=a , default='''summarization''' , help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' , type=a , default=8 , required=a , help='''batch size''' ) parser.add_argument( '''--n_obs''' , type=a , default=-1 , required=a , help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' , action='''store_true''' ) parser.add_argument('''--dump-args''' , action='''store_true''' , help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' , nargs='''?''' , type=a , const=datetime_now() , help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) , ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate _A , _A: Tuple = parser.parse_known_args() _A: List[str] = parse_numeric_n_bool_cl_kwargs(a ) if parsed_args and verbose: print(f"""parsed the following generate kwargs: {parsed_args}""" ) _A: int = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: _A: List[str] = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=a ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(f"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) _A: Dict = generate_summaries_or_translations( a , args.save_path , args.model_name , batch_size=args.bs , device=args.device , fpaa=args.fpaa , task=args.task , prefix=args.prefix , **a , ) if args.reference_path is None: return {} # Compute scores _A: Dict = calculate_bleu if '''translation''' in args.task else calculate_rouge _A: List[Any] = [x.rstrip() for x in open(args.save_path ).readlines()] _A: Any = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(a )] _A: dict = score_fn(a , a ) scores.update(a ) if args.dump_args: scores.update(a ) if args.info: _A: Optional[Any] = args.info if verbose: print(a ) if args.score_path is not None: json.dump(a , open(args.score_path , '''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
301
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : Union[str, Any] = {'configuration_mmbt': ['MMBTConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Optional[Any] = ['MMBTForClassification', 'MMBTModel', 'ModalEmbeddings'] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys __lowerCAmelCase : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
88
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : Any = { 'configuration_mctct': ['MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MCTCTConfig'], 'feature_extraction_mctct': ['MCTCTFeatureExtractor'], 'processing_mctct': ['MCTCTProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : int = [ 'MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MCTCTForCTC', 'MCTCTModel', 'MCTCTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __lowerCAmelCase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
88
1
"""simple docstring""" import os import sys UpperCamelCase_ = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) UpperCamelCase_ = [ 'torch', 'numpy', 'tokenizers', 'filelock', 'requests', 'tqdm', 'regex', 'sentencepiece', 'sacremoses', 'importlib_metadata', 'huggingface_hub', ] @add_start_docstrings(AutoConfig.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Union[str, Any]: """simple docstring""" return AutoConfig.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Tuple: """simple docstring""" return AutoTokenizer.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Dict: """simple docstring""" return AutoModel.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Any: """simple docstring""" return AutoModelForCausalLM.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Union[str, Any]: """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Optional[Any]: """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*UpperCAmelCase , **UpperCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def UpperCamelCase ( *UpperCAmelCase , **UpperCAmelCase ) ->Optional[Any]: """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*UpperCAmelCase , **UpperCAmelCase )
303
"""simple docstring""" import os import numpy import onnx def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ) ->List[str]: """simple docstring""" a_ = a.name a_ = b.name a_ = "" a_ = "" a_ = a == b a_ = name_a a_ = name_b return res def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->List[Any]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(UpperCAmelCase , UpperCAmelCase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , UpperCAmelCase , UpperCAmelCase ) _graph_replace_input_with(node_proto.attribute[1].g , UpperCAmelCase , UpperCAmelCase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->Dict: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->int: """simple docstring""" a_ = list(model.graph.initializer ) a_ = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i a_ = inits[i].name a_ = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase ) ->Union[str, Any]: """simple docstring""" a_ = os.path.dirname(UpperCAmelCase ) a_ = os.path.basename(UpperCAmelCase ) a_ = onnx.load(os.path.join(UpperCAmelCase , UpperCAmelCase ) ) a_ = list(model.graph.initializer ) a_ = set() a_ = {} a_ = [] a_ = 0 for i in range(len(UpperCAmelCase ) ): if i in dup_set: continue for j in range(i + 1 , len(UpperCAmelCase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(UpperCAmelCase ) dup_set.add(UpperCAmelCase ) a_ = inits[j].data_type a_ = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print("unexpected data type: " , UpperCAmelCase ) total_reduced_size += mem_size a_ = inits[i].name a_ = inits[j].name if name_i in dup_map: dup_map[name_i].append(UpperCAmelCase ) else: a_ = [name_j] ind_to_replace.append((j, i) ) print("total reduced size: " , total_reduced_size / 1_024 / 1_024 / 1_024 , "GB" ) a_ = sorted(UpperCAmelCase ) _remove_dup_initializers_from_model(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) a_ = "optimized_" + model_file_name a_ = os.path.join(UpperCAmelCase , UpperCAmelCase ) onnx.save(UpperCAmelCase , UpperCAmelCase ) return new_model
303
1
import numpy as np def lowerCAmelCase_ ( _snake_case : np.ndarray , _snake_case : float ) -> np.ndarray: '''simple docstring''' return np.where(vector > 0 , _snake_case , (alpha * (np.exp(_snake_case ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
281
import math def lowerCAmelCase_ ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' return math.pow(_snake_case , 2 ) - a def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' return 2 * x def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' __magic_name__ : Optional[int] = 2.0 while start <= a: __magic_name__ : str = math.pow(_snake_case , 2 ) return start def lowerCAmelCase_ ( _snake_case : float , _snake_case : int = 9999 , _snake_case : float = 0.00_000_000_000_001 ) -> float: '''simple docstring''' if a < 0: raise ValueError("math domain error" ) __magic_name__ : Optional[int] = get_initial_point(_snake_case ) for _ in range(_snake_case ): __magic_name__ : int = value __magic_name__ : str = value - fx(_snake_case , _snake_case ) / fx_derivative(_snake_case ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
281
1
'''simple docstring''' from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup __UpperCAmelCase : Tuple = '''https://www.indeed.co.in/jobs?q=mobile+app+development&l=''' def __A ( lowerCAmelCase_ = "mumbai" ): _UpperCAmelCase : Optional[Any] = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): _UpperCAmelCase : str = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() _UpperCAmelCase : Union[str, Any] = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs('''Bangalore'''), 1): print(F"Job {i:>2} is {job[0]} at {job[1]}")
367
'''simple docstring''' def __A ( lowerCAmelCase_ ): _UpperCAmelCase : Optional[Any] = 0 while len(lowerCAmelCase_ ) > 1: _UpperCAmelCase : List[Any] = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): _UpperCAmelCase : Optional[Any] = files.index(min(lowerCAmelCase_ ) ) temp += files[min_index] files.pop(lowerCAmelCase_ ) files.append(lowerCAmelCase_ ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
170
0
from math import factorial UpperCAmelCase = {str(digit): factorial(digit) for digit in range(10)} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): raise TypeError('Parameter number must be int' ) if number < 0: raise ValueError('Parameter number must be greater than or equal to 0' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(__SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 60 , __SCREAMING_SNAKE_CASE = 100_0000 ): if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) or not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): raise TypeError('Parameters chain_length and number_limit must be int' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( 'Parameters chain_length and number_limit must be greater than 0' ) # the counter for the chains with the exact desired length lowercase = 0 # the cached sizes of the previous chains lowercase = {} for start_chain_element in range(1 , __SCREAMING_SNAKE_CASE ): # The temporary set will contain the elements of the chain lowercase = set() lowercase = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. lowercase = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(__SCREAMING_SNAKE_CASE ) chain_set_length += 1 lowercase = digit_factorial_sum(__SCREAMING_SNAKE_CASE ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] lowercase = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
195
import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient UpperCAmelCase = WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = test_results.split(' ' ) lowercase = 0 lowercase = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. lowercase = expressions[-2] if '=' in expressions[-1] else expressions[-1] for i, expression in enumerate(__SCREAMING_SNAKE_CASE ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = {} lowercase = None lowercase = False for line in failures_short_lines.split('\n' ): if re.search(r'_ \[doctest\]' , __SCREAMING_SNAKE_CASE ): lowercase = True lowercase = line.split(' ' )[2] elif in_error and not line.split(' ' )[0].isdigit(): lowercase = line lowercase = False return failures class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case ): lowercase = title lowercase = doc_test_results['time_spent'].split(',' )[0] lowercase = doc_test_results['success'] lowercase = doc_test_results['failures'] lowercase = self.n_success + self.n_failures # Failures and success of the modeling tests lowercase = doc_test_results @property def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [self._time_spent] lowercase = 0 for time in time_spent: lowercase = time.split(':' ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(snake_case ) == 1: lowercase = [0, 0, time_parts[0]] lowercase , lowercase , lowercase = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3600 + minutes * 60 + seconds lowercase , lowercase , lowercase = total_secs // 3600, (total_secs % 3600) // 60, total_secs % 60 return F'''{int(snake_case )}h{int(snake_case )}m{int(snake_case )}s''' @property def SCREAMING_SNAKE_CASE__ ( self ): return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def SCREAMING_SNAKE_CASE__ ( self ): return { "type": "section", "text": { "type": "plain_text", "text": F'''🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.''', "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def SCREAMING_SNAKE_CASE__ ( self ): return { "type": "section", "text": { "type": "plain_text", "text": ( F'''There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in''' F''' {self.time}.''' ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def SCREAMING_SNAKE_CASE__ ( self ): lowercase = 40 lowercase = {k: v['failed'] for k, v in doc_test_results.items() if isinstance(snake_case , snake_case )} lowercase = '' for category, failures in category_failures.items(): if len(snake_case ) == 0: continue if report != "": report += "\n\n" report += F'''*{category} failures*:'''.ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(snake_case ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": F'''The following examples had failures:\n\n\n{report}\n''', }, } @property def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(snake_case ) @staticmethod def SCREAMING_SNAKE_CASE__ ( ): lowercase = [ { 'type': 'section', 'text': { 'type': 'plain_text', 'text': 'There was an issue running the tests.', }, 'accessory': { 'type': 'button', 'text': {'type': 'plain_text', 'text': 'Check Action results', 'emoji': True}, 'url': F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } ] print('Sending the following payload' ) print(json.dumps({'blocks': json.loads(snake_case )} ) ) client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , text='There was an issue running the tests.' , blocks=snake_case , ) def SCREAMING_SNAKE_CASE__ ( self ): print('Sending the following payload' ) print(json.dumps({'blocks': json.loads(self.payload )} ) ) lowercase = F'''{self.n_failures} failures out of {self.n_tests} tests,''' if self.n_failures else 'All tests passed.' lowercase = client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , blocks=self.payload , text=snake_case , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): lowercase = '' for key, value in failures.items(): lowercase = value[:200] + ' [Truncated]' if len(snake_case ) > 250 else value failures_text += F'''*{key}*\n_{value}_\n\n''' lowercase = job_name lowercase = {'type': 'section', 'text': {'type': 'mrkdwn', 'text': text}} if job_link is not None: lowercase = { 'type': 'button', 'text': {'type': 'plain_text', 'text': 'GitHub Action job', 'emoji': True}, 'url': job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def SCREAMING_SNAKE_CASE__ ( self ): if self.thread_ts is None: raise ValueError('Can only post reply if a post has been made.' ) lowercase = self.doc_test_results.pop('job_link' ) self.doc_test_results.pop('failures' ) self.doc_test_results.pop('success' ) self.doc_test_results.pop('time_spent' ) lowercase = sorted(self.doc_test_results.items() , key=lambda snake_case : t[0] ) for job, job_result in sorted_dict: if len(job_result['failures'] ): lowercase = F'''*Num failures* :{len(job_result['failed'] )} \n''' lowercase = job_result['failures'] lowercase = self.get_reply_blocks(snake_case , snake_case , snake_case , text=snake_case ) print('Sending the following reply' ) print(json.dumps({'blocks': blocks} ) ) client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , text=F'''Results for {job}''' , blocks=snake_case , thread_ts=self.thread_ts['ts'] , ) time.sleep(1 ) def UpperCAmelCase_ ( ): lowercase = os.environ['GITHUB_RUN_ID'] lowercase = F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100''' lowercase = requests.get(__SCREAMING_SNAKE_CASE ).json() lowercase = {} try: jobs.update({job['name']: job['html_url'] for job in result['jobs']} ) lowercase = math.ceil((result['total_count'] - 100) / 100 ) for i in range(__SCREAMING_SNAKE_CASE ): lowercase = requests.get(url + F'''&page={i + 2}''' ).json() jobs.update({job['name']: job['html_url'] for job in result['jobs']} ) return jobs except Exception as e: print('Unknown error, could not fetch links.' , __SCREAMING_SNAKE_CASE ) return {} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = {} if os.path.exists(__SCREAMING_SNAKE_CASE ): lowercase = os.listdir(__SCREAMING_SNAKE_CASE ) for file in files: try: with open(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , encoding='utf-8' ) as f: lowercase = f.read() except UnicodeDecodeError as e: raise ValueError(F'''Could not open {os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )}.''' ) from e return _artifact def UpperCAmelCase_ ( ): class A_ : '''simple docstring''' def __init__( self , snake_case ): lowercase = name lowercase = [] def __str__( self ): return self.name def SCREAMING_SNAKE_CASE__ ( self , snake_case ): self.paths.append({'name': self.name, 'path': path} ) lowercase = {} lowercase = filter(os.path.isdir , os.listdir() ) for directory in directories: lowercase = directory if artifact_name not in _available_artifacts: lowercase = Artifact(__SCREAMING_SNAKE_CASE ) _available_artifacts[artifact_name].add_path(__SCREAMING_SNAKE_CASE ) return _available_artifacts if __name__ == "__main__": UpperCAmelCase = get_job_links() UpperCAmelCase = retrieve_available_artifacts() UpperCAmelCase = collections.OrderedDict( [ ('''*.py''', '''API Examples'''), ('''*.md''', '''MD Examples'''), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' UpperCAmelCase = { v: { '''failed''': [], '''failures''': {}, } for v in docs.values() } # Link to the GitHub Action job UpperCAmelCase = github_actions_job_links.get('''run_doctests''') UpperCAmelCase = available_artifacts['''doc_tests_gpu_test_reports'''].paths[0] UpperCAmelCase = retrieve_artifact(artifact_path['''name''']) if "stats" in artifact: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = handle_test_results(artifact['''stats''']) UpperCAmelCase = failed UpperCAmelCase = success UpperCAmelCase = time_spent[1:-1] + ''', ''' UpperCAmelCase = extract_first_line_failure(artifact['''failures_short''']) for line in artifact["summary_short"].split('''\n'''): if re.search('''FAILED''', line): UpperCAmelCase = line.replace('''FAILED ''', '''''') UpperCAmelCase = line.split()[0].replace('''\n''', '''''') if "::" in line: UpperCAmelCase , UpperCAmelCase = line.split('''::''') else: UpperCAmelCase , UpperCAmelCase = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): UpperCAmelCase = docs[file_regex] doc_test_results[category]["failed"].append(test) UpperCAmelCase = all_failures[test] if test in all_failures else '''N/A''' UpperCAmelCase = failure break UpperCAmelCase = Message('''🤗 Results of the doc tests.''', doc_test_results) message.post() message.post_reply()
195
1
'''simple docstring''' from __future__ import annotations import math def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if not scores: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) , minimax(depth + 1 , node_index * 2 + 1 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) , minimax(depth + 1 , node_index * 2 + 1 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) , ) ) def UpperCamelCase( ): UpperCAmelCase : Dict = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] UpperCAmelCase : Optional[int] = math.log(len(UpperCAmelCase_ ) , 2 ) print(F"""Optimal value : {minimax(0 , 0 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
353
'''simple docstring''' from __future__ import annotations def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if len(UpperCAmelCase_ ) == 0: raise ValueError('find_max() arg is an empty sequence' ) if ( left >= len(UpperCAmelCase_ ) or left < -len(UpperCAmelCase_ ) or right >= len(UpperCAmelCase_ ) or right < -len(UpperCAmelCase_ ) ): raise IndexError('list index out of range' ) if left == right: return nums[left] UpperCAmelCase : Optional[int] = (left + right) >> 1 # the middle UpperCAmelCase : Any = find_max(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # find max in range[left, mid] UpperCAmelCase : Union[str, Any] = find_max(UpperCAmelCase_ , mid + 1 , UpperCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
280
0
import numpy as np def lowerCamelCase__ ( A__ : np.ndarray ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def lowerCamelCase__ ( A__ : np.ndarray ): '''simple docstring''' return vector * sigmoid(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
12
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Any , UpperCamelCase_: str , UpperCamelCase_: Dict ): __lowerCamelCase = question_encoder __lowerCamelCase = generator __lowerCamelCase = self.question_encoder def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[Any] ): if os.path.isfile(UpperCamelCase_ ): raise ValueError(F'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(UpperCamelCase_ , exist_ok=UpperCamelCase_ ) __lowerCamelCase = os.path.join(UpperCamelCase_ , """question_encoder_tokenizer""" ) __lowerCamelCase = os.path.join(UpperCamelCase_ , """generator_tokenizer""" ) self.question_encoder.save_pretrained(UpperCamelCase_ ) self.generator.save_pretrained(UpperCamelCase_ ) @classmethod def lowerCAmelCase__ ( cls: List[Any] , UpperCamelCase_: Dict , **UpperCamelCase_: Union[str, Any] ): # dynamically import AutoTokenizer from ..auto.tokenization_auto import AutoTokenizer __lowerCamelCase = kwargs.pop("""config""" , UpperCamelCase_ ) if config is None: __lowerCamelCase = RagConfig.from_pretrained(UpperCamelCase_ ) __lowerCamelCase = AutoTokenizer.from_pretrained( UpperCamelCase_ , config=config.question_encoder , subfolder="""question_encoder_tokenizer""" ) __lowerCamelCase = AutoTokenizer.from_pretrained( UpperCamelCase_ , config=config.generator , subfolder="""generator_tokenizer""" ) return cls(question_encoder=UpperCamelCase_ , generator=UpperCamelCase_ ) def __call__( self: Tuple , *UpperCamelCase_: int , **UpperCamelCase_: int ): return self.current_tokenizer(*UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple , *UpperCamelCase_: List[Any] , **UpperCamelCase_: List[Any] ): return self.generator.batch_decode(*UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] , *UpperCamelCase_: str , **UpperCamelCase_: Union[str, Any] ): return self.generator.decode(*UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.question_encoder def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.generator def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[List[str]] = None , UpperCamelCase_: Optional[int] = None , UpperCamelCase_: Optional[int] = None , UpperCamelCase_: str = "longest" , UpperCamelCase_: str = None , UpperCamelCase_: bool = True , **UpperCamelCase_: int , ): warnings.warn( """`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the """ """regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` """ """context manager to prepare your targets. See the documentation of your specific tokenizer for more """ """details""" , UpperCamelCase_ , ) if max_length is None: __lowerCamelCase = self.current_tokenizer.model_max_length __lowerCamelCase = self( UpperCamelCase_ , add_special_tokens=UpperCamelCase_ , return_tensors=UpperCamelCase_ , max_length=UpperCamelCase_ , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , **UpperCamelCase_ , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: __lowerCamelCase = self.current_tokenizer.model_max_length __lowerCamelCase = self( text_target=UpperCamelCase_ , add_special_tokens=UpperCamelCase_ , return_tensors=UpperCamelCase_ , padding=UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_ , **UpperCamelCase_ , ) __lowerCamelCase = labels["""input_ids"""] return model_inputs
12
1
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A__ = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class a ( __lowerCamelCase , unittest.TestCase ): __lowerCAmelCase : Tuple = XGLMTokenizer __lowerCAmelCase : Optional[int] = XGLMTokenizerFast __lowerCAmelCase : Any = True __lowerCAmelCase : int = True def __lowerCamelCase ( self :Any ): super().setUp() # We have a SentencePiece fixture for testing snake_case__ : Optional[int] = XGLMTokenizer(__lowercase ,keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCamelCase ( self :int ): snake_case__ : List[str] = '''<pad>''' snake_case__ : Union[str, Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) ,__lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) ,__lowercase ) def __lowerCamelCase ( self :Optional[Any] ): snake_case__ : List[str] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'''<s>''' ) self.assertEqual(vocab_keys[1] ,'''<pad>''' ) self.assertEqual(len(__lowercase ) ,1_0_0_8 ) def __lowerCamelCase ( self :Optional[int] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_0_0_8 ) def __lowerCamelCase ( self :Any ): snake_case__ : Optional[int] = XGLMTokenizer(__lowercase ,keep_accents=__lowercase ) snake_case__ : Tuple = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase ,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) ,[value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] ,) snake_case__ : Optional[int] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] ,) snake_case__ : str = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase ,[ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] ,) snake_case__ : Dict = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] ,) @cached_property def __lowerCamelCase ( self :List[str] ): return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def __lowerCamelCase ( self :Tuple ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__lowercase ,f.name ) snake_case__ : Any = XGLMTokenizer(f.name ,keep_accents=__lowercase ) snake_case__ : int = pickle.dumps(__lowercase ) pickle.loads(__lowercase ) def __lowerCamelCase ( self :Any ): if not self.test_rust_tokenizer: return snake_case__ : List[Any] = self.get_tokenizer() snake_case__ : Dict = self.get_rust_tokenizer() snake_case__ : Optional[int] = '''I was born in 92000, and this is falsé.''' snake_case__ : int = tokenizer.tokenize(__lowercase ) snake_case__ : Union[str, Any] = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase ,__lowercase ) snake_case__ : Any = tokenizer.encode(__lowercase ,add_special_tokens=__lowercase ) snake_case__ : Optional[int] = rust_tokenizer.encode(__lowercase ,add_special_tokens=__lowercase ) self.assertListEqual(__lowercase ,__lowercase ) snake_case__ : List[str] = self.get_rust_tokenizer() snake_case__ : Dict = tokenizer.encode(__lowercase ) snake_case__ : Optional[Any] = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase ,__lowercase ) @slow def __lowerCamelCase ( self :str ): snake_case__ : Any = '''Hello World!''' snake_case__ : List[str] = [2, 3_1_2_2_7, 4_4_4_7, 3_5] self.assertListEqual(__lowercase ,self.big_tokenizer.encode(__lowercase ) ) @slow def __lowerCamelCase ( self :Any ): snake_case__ : Any = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off snake_case__ : List[Any] = [2, 1_0_1_8, 6_7, 1_1, 1_9_8_8, 2_6_1_7, 5_6_3_1, 2_7_8, 1_1, 3_4_0_7, 4_8, 7_1_6_3_0, 2_8_0_8_5, 4, 3_2_3_4, 1_5_7, 1_3, 6, 5, 6, 4, 3_5_2_6, 7_6_8, 1_5, 6_5_9, 5_7, 2_9_8, 3_9_8_3, 8_6_4, 1_2_9, 2_1, 6, 5, 1_3_6_7_5, 3_7_7, 6_5_2, 7_5_8_0, 1_0_3_4_1, 1_5_5, 2_8_1_7, 4_2_2, 1_6_6_6, 7, 1_6_7_4, 5_3, 1_1_3, 2_0_2_2_7_7, 1_7_8_9_2, 3_3, 6_0, 8_7, 4, 3_2_3_4, 1_5_7, 6_1, 2_6_6_7, 5_2_3_7_6, 1_9, 8_8, 2_3, 7_3_5] # fmt: on self.assertListEqual(__lowercase ,self.big_tokenizer.encode(__lowercase ) ) @slow def __lowerCamelCase ( self :str ): # fmt: off snake_case__ : Any = { '''input_ids''': [[2, 1_0_8_8_2_5, 1_1_6_3, 1_5, 8_8_0_1_0, 4_7_3, 1_5_8_9_8, 1_5_7, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 2_3_8_0_2_1, 1_1_6_3, 5_3, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 5_3_2_8_3, 1_8_2_3_9_6, 8, 1_8_5_6_6, 1_6, 3_6_7_3_3, 4_1_0_1, 8, 2_3_0, 2_4_4_0_1_7, 1_2_2_5_5_3, 7, 1_5, 1_3_2_5_9_7, 4, 2_9_3, 1_2_5_1_1, 7_6_1_0, 4, 3_4_1_4, 1_3_2_5_9_7, 9, 4, 3_2_3_6_1, 3_6_2, 4, 7_3_4, 2_8_5_1_2, 3_2_5_6_9, 1_8, 4, 3_2_3_6_1, 2_6_0_9_6, 1_4_9_8_2, 7_3, 1_8_7_1_5, 2_1_4_3_3, 2_3_5_2_6_1, 1_5, 4_9_2, 1_2_4_2_7, 1_6, 5_3, 1_8_7_1_5, 2_1_4_3_3, 6_5_4_5_4, 1_5, 2_3_6_5_9, 5_6_3, 1_6, 2_7_8, 5_9_7, 2_8_4_3, 5_9_5, 7_9_3_1, 1_8_2_3_9_6, 6_4_1_8_6, 2_2, 8_8_6, 5_9_5, 1_3_2_9_8_1, 5_3, 2_5_5_4_0, 3_4_4_9, 4_3_9_8_2, 3_9_9_0_1, 5_9_5_1, 8_7_8, 3_3_0, 4, 2_7_6_9_4, 8_0_2_6_9, 3_1_2, 5_3, 6_5_1_7, 1_1_7_8_0, 6_1_1, 2_0_4_0_8, 5], [2, 6, 1_3_2_5_9_7, 6_7, 4_2_8_9_7, 3_3, 5_9_2, 8, 1_6_3_7_2_9, 2_5_5_4_0, 3_6_1, 1_3_6_9_9_7, 1_0_9_5_1_4, 1_7_3_2_3_0, 7, 5_0_1, 6_0, 1_0_2_9_1_3, 1_9_6, 5_6_3_1, 2_3_5, 6_3_2_4_3, 4_7_3, 6, 2_3_1_7_5_7, 7_4, 5_2_7_7, 7_9_0_5, 5_3, 3_0_9_5, 3_7_3_1_7, 2_2, 4_5_4, 1_8_3_8_7_4, 5], [2, 2_6_8, 3_1_2_9_8, 4_6_5_3_0, 6, 1_3_2_9_3_5, 4_3_8_3_1, 7, 5_9_7, 3_2, 2_4, 3_6_8_8, 9_8_6_5, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowercase ,model_name='''facebook/xglm-564M''' ,padding=__lowercase ,)
44
A__ = 256 # Modulus to hash a string A__ = 100_0003 def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> bool: """simple docstring""" snake_case__ : str = len(__lowerCAmelCase ) snake_case__ : Optional[int] = len(__lowerCAmelCase ) if p_len > t_len: return False snake_case__ : str = 0 snake_case__ : Union[str, Any] = 0 snake_case__ : Dict = 1 # Calculating the hash of pattern and substring of text for i in range(__lowerCAmelCase ): snake_case__ : int = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus snake_case__ : str = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue snake_case__ : str = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash snake_case__ : Any = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def _lowerCAmelCase ( ) -> None: """simple docstring""" snake_case__ : Optional[int] = '''abc1abc12''' snake_case__ : Dict = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' snake_case__ : int = '''alskfjaldsk23adsfabcabc''' assert rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) and not rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) # Test 2) snake_case__ : int = '''ABABX''' snake_case__ : Any = '''ABABZABABYABABX''' assert rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) # Test 3) snake_case__ : Dict = '''AAAB''' snake_case__ : Union[str, Any] = '''ABAAAAAB''' assert rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) # Test 4) snake_case__ : Union[str, Any] = '''abcdabcy''' snake_case__ : Optional[Any] = '''abcxabcdabxabcdabcdabcy''' assert rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) # Test 5) snake_case__ : Dict = '''Lü''' snake_case__ : Optional[Any] = '''Lüsai''' assert rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : str = '''Lue''' assert not rabin_karp(__lowerCAmelCase , __lowerCAmelCase ) print('''Success.''' ) if __name__ == "__main__": test_rabin_karp()
44
1
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def _UpperCamelCase ( snake_case__, snake_case__, snake_case__, snake_case__, snake_case__ ) -> np.ndarray: __UpperCAmelCase : int = cva.getAffineTransform(snake_case__, snake_case__ ) return cva.warpAffine(snake_case__, snake_case__, (rows, cols) ) if __name__ == "__main__": # read original image _snake_case = cva.imread( str(Path(__file__).resolve().parent.parent / '''image_data''' / '''lena.jpg''') ) # turn image in gray scale value _snake_case = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape _snake_case , _snake_case = gray_img.shape # set different points to rotate image _snake_case = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) _snake_case = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) _snake_case = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) _snake_case = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list _snake_case = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations _snake_case = plt.figure(1) _snake_case = ['''Original''', '''Rotation 1''', '''Rotation 2''', '''Rotation 3'''] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, '''gray''') plt.title(titles[i]) plt.axis('''off''') plt.subplots_adjust(left=0.0, bottom=0.0_5, right=1.0, top=0.9_5) plt.show()
157
import argparse import os import re _snake_case = '''src/transformers/models/auto''' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict _snake_case = re.compile(r'''[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict''') # re pattern that matches identifiers in mappings _snake_case = re.compile(r'''\s*\(\s*"(\S[^"]+)"''') def _UpperCamelCase ( snake_case__, snake_case__ = False ) -> List[Any]: with open(snake_case__, "r", encoding="utf-8" ) as f: __UpperCAmelCase : Dict = f.read() __UpperCAmelCase : Optional[Any] = content.split("\n" ) __UpperCAmelCase : int = [] __UpperCAmelCase : Optional[int] = 0 while line_idx < len(snake_case__ ): if _re_intro_mapping.search(lines[line_idx] ) is not None: __UpperCAmelCase : str = len(re.search(r"^(\s*)\S", lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(" " * indent + "(" ): new_lines.append(lines[line_idx] ) line_idx += 1 __UpperCAmelCase : Dict = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": __UpperCAmelCase : str = line_idx while not lines[line_idx].startswith(" " * indent + ")" ): line_idx += 1 blocks.append("\n".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers __UpperCAmelCase : Dict = sorted(snake_case__, key=lambda snake_case__ : _re_identifier.search(snake_case__ ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(snake_case__, "w", encoding="utf-8" ) as f: f.write("\n".join(snake_case__ ) ) elif "\n".join(snake_case__ ) != content: return True def _UpperCamelCase ( snake_case__ = False ) -> Any: __UpperCAmelCase : str = [os.path.join(snake_case__, snake_case__ ) for f in os.listdir(snake_case__ ) if f.endswith(".py" )] __UpperCAmelCase : Optional[Any] = [sort_auto_mapping(snake_case__, overwrite=snake_case__ ) for fname in fnames] if not overwrite and any(snake_case__ ): __UpperCAmelCase : List[Any] = [f for f, d in zip(snake_case__, snake_case__ ) if d] raise ValueError( f'''The following files have auto mappings that need sorting: {', '.join(snake_case__ )}. Run `make style` to fix''' " this." ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') _snake_case = parser.parse_args() sort_all_auto_mappings(not args.check_only)
157
1
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=A ) class lowercase_ ( A ): """simple docstring""" lowerCamelCase_ = field(default='''summarization''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowerCamelCase_ = Features({'''text''': Value('''string''' )} ) lowerCamelCase_ = Features({'''summary''': Value('''string''' )} ) lowerCamelCase_ = '''text''' lowerCamelCase_ = '''summary''' @property def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
351
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( __A : list ) -> list: if len(__A ) <= 1: return lst _SCREAMING_SNAKE_CASE = 1 while i < len(__A ): if lst[i - 1] <= lst[i]: i += 1 else: _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = lst[i], lst[i - 1] i -= 1 if i == 0: _SCREAMING_SNAKE_CASE = 1 return lst if __name__ == "__main__": lowerCamelCase_ = input('Enter numbers separated by a comma:\n').strip() lowerCamelCase_ = [int(item) for item in user_input.split(',')] print(gnome_sort(unsorted))
111
0
"""simple docstring""" import os import sys __A = os.path.join(os.path.dirname(__file__), "src") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __A = [ "torch", "numpy", "tokenizers", "filelock", "requests", "tqdm", "regex", "sentencepiece", "sacremoses", "importlib_metadata", "huggingface_hub", ] @add_start_docstrings(AutoConfig.__doc__ ) def UpperCamelCase__ ( *lowercase__ : Optional[int] , **lowercase__ : Optional[int] ): return AutoConfig.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def UpperCamelCase__ ( *lowercase__ : int , **lowercase__ : List[str] ): return AutoTokenizer.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoModel.__doc__ ) def UpperCamelCase__ ( *lowercase__ : str , **lowercase__ : Dict ): return AutoModel.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def UpperCamelCase__ ( *lowercase__ : int , **lowercase__ : Optional[int] ): return AutoModelForCausalLM.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def UpperCamelCase__ ( *lowercase__ : Tuple , **lowercase__ : Union[str, Any] ): return AutoModelForMaskedLM.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def UpperCamelCase__ ( *lowercase__ : str , **lowercase__ : str ): return AutoModelForSequenceClassification.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def UpperCamelCase__ ( *lowercase__ : Tuple , **lowercase__ : Optional[int] ): return AutoModelForQuestionAnswering.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ )
148
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowercase ( unittest.TestCase ): def A__ ( self): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def A__ ( self): lowercase = 1 lowercase = 3 lowercase = (3_2, 3_2) lowercase = floats_tensor((batch_size, num_channels) + sizes ,rng=random.Random(0)).to(A__) return image @property def A__ ( self): torch.manual_seed(0) lowercase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) ,layers_per_block=2 ,sample_size=3_2 ,in_channels=4 ,out_channels=4 ,down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') ,up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') ,cross_attention_dim=3_2 ,) return model @property def A__ ( self): torch.manual_seed(0) lowercase = AutoencoderKL( block_out_channels=[3_2, 6_4] ,in_channels=3 ,out_channels=3 ,down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] ,up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] ,latent_channels=4 ,) return model @property def A__ ( self): torch.manual_seed(0) lowercase = RobertaSeriesConfig( hidden_size=3_2 ,project_dim=3_2 ,intermediate_size=3_7 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=5_0_0_6 ,) return RobertaSeriesModelWithTransformation(A__) @property def A__ ( self): def extract(*A__ ,**A__): class lowercase : def __init__( self): lowercase = torch.ones([0]) def A__ ( self ,A__): self.pixel_values.to(A__) return self return Out() return extract def A__ ( self): lowercase = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowercase = self.dummy_cond_unet lowercase = PNDMScheduler(skip_prk_steps=A__) lowercase = self.dummy_vae lowercase = self.dummy_text_encoder lowercase = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''') lowercase = 7_7 lowercase = self.dummy_image.to(A__) lowercase = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk lowercase = AltDiffusionImgaImgPipeline( unet=A__ ,scheduler=A__ ,vae=A__ ,text_encoder=A__ ,tokenizer=A__ ,safety_checker=A__ ,feature_extractor=self.dummy_extractor ,) lowercase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor ,do_normalize=A__) lowercase = alt_pipe.to(A__) alt_pipe.set_progress_bar_config(disable=A__) lowercase = '''A painting of a squirrel eating a burger''' lowercase = torch.Generator(device=A__).manual_seed(0) lowercase = alt_pipe( [prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type='''np''' ,image=A__ ,) lowercase = output.images lowercase = torch.Generator(device=A__).manual_seed(0) lowercase = alt_pipe( [prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type='''np''' ,image=A__ ,return_dict=A__ ,)[0] lowercase = image[0, -3:, -3:, -1] lowercase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowercase = np.array([0.4427, 0.3731, 0.4249, 0.4941, 0.4546, 0.4148, 0.4193, 0.4666, 0.4499]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 5E-3 @unittest.skipIf(torch_device != '''cuda''' ,'''This test requires a GPU''') def A__ ( self): lowercase = self.dummy_cond_unet lowercase = PNDMScheduler(skip_prk_steps=A__) lowercase = self.dummy_vae lowercase = self.dummy_text_encoder lowercase = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''') lowercase = 7_7 lowercase = self.dummy_image.to(A__) # put models in fp16 lowercase = unet.half() lowercase = vae.half() lowercase = bert.half() # make sure here that pndm scheduler skips prk lowercase = AltDiffusionImgaImgPipeline( unet=A__ ,scheduler=A__ ,vae=A__ ,text_encoder=A__ ,tokenizer=A__ ,safety_checker=A__ ,feature_extractor=self.dummy_extractor ,) lowercase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor ,do_normalize=A__) lowercase = alt_pipe.to(A__) alt_pipe.set_progress_bar_config(disable=A__) lowercase = '''A painting of a squirrel eating a burger''' lowercase = torch.manual_seed(0) lowercase = alt_pipe( [prompt] ,generator=A__ ,num_inference_steps=2 ,output_type='''np''' ,image=A__ ,).images assert image.shape == (1, 3_2, 3_2, 3) @unittest.skipIf(torch_device != '''cuda''' ,'''This test requires a GPU''') def A__ ( self): lowercase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''') # resize to resolution that is divisible by 8 but not 16 or 32 lowercase = init_image.resize((7_6_0, 5_0_4)) lowercase = '''BAAI/AltDiffusion''' lowercase = AltDiffusionImgaImgPipeline.from_pretrained( A__ ,safety_checker=A__ ,) pipe.to(A__) pipe.set_progress_bar_config(disable=A__) pipe.enable_attention_slicing() lowercase = '''A fantasy landscape, trending on artstation''' lowercase = torch.manual_seed(0) lowercase = pipe( prompt=A__ ,image=A__ ,strength=0.75 ,guidance_scale=7.5 ,generator=A__ ,output_type='''np''' ,) lowercase = output.images[0] lowercase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 7_6_0, 3) lowercase = np.array([0.9358, 0.9397, 0.9599, 0.9901, 1.0000, 1.0000, 0.9882, 1.0000, 1.0000]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch_gpu class lowercase ( unittest.TestCase ): def A__ ( self): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self): lowercase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''') lowercase = init_image.resize((7_6_8, 5_1_2)) lowercase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''') lowercase = '''BAAI/AltDiffusion''' lowercase = AltDiffusionImgaImgPipeline.from_pretrained( A__ ,safety_checker=A__ ,) pipe.to(A__) pipe.set_progress_bar_config(disable=A__) pipe.enable_attention_slicing() lowercase = '''A fantasy landscape, trending on artstation''' lowercase = torch.manual_seed(0) lowercase = pipe( prompt=A__ ,image=A__ ,strength=0.75 ,guidance_scale=7.5 ,generator=A__ ,output_type='''np''' ,) lowercase = output.images[0] assert image.shape == (5_1_2, 7_6_8, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image).max() < 1E-2
101
0
"""simple docstring""" from PIL import Image def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase , __lowerCAmelCase : List[Any] = image.size __lowerCAmelCase : Optional[Any] = 0 __lowerCAmelCase : Tuple = image.load() for i in range(_UpperCamelCase ): for j in range(_UpperCamelCase ): __lowerCAmelCase : int = pixels[j, i] mean += pixel mean //= width * height for j in range(_UpperCamelCase ): for i in range(_UpperCamelCase ): __lowerCAmelCase : Optional[Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowerCamelCase__ = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
182
"""simple docstring""" import os def __lowerCAmelCase (_UpperCamelCase = "input.txt" ): with open(os.path.join(os.path.dirname(_UpperCamelCase ) , _UpperCamelCase ) ) as input_file: __lowerCAmelCase : Optional[Any] = [ [int(_UpperCamelCase ) for element in line.split(',' )] for line in input_file.readlines() ] __lowerCAmelCase : List[str] = len(_UpperCamelCase ) __lowerCAmelCase : Tuple = len(matrix[0] ) __lowerCAmelCase : int = [[-1 for _ in range(_UpperCamelCase )] for _ in range(_UpperCamelCase )] for i in range(_UpperCamelCase ): __lowerCAmelCase : Any = matrix[i][0] for j in range(1 , _UpperCamelCase ): for i in range(_UpperCamelCase ): __lowerCAmelCase : Optional[Any] = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , _UpperCamelCase ): __lowerCAmelCase : Optional[Any] = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): __lowerCAmelCase : List[str] = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(f'{solution() = }')
182
1
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ = 100 ): __SCREAMING_SNAKE_CASE = set() __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = n + 1 # maximum limit for a in range(2 , UpperCamelCase_ ): for b in range(2 , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = a**b # calculates the current power collect_powers.add(UpperCamelCase_ ) # adds the result to the set return len(UpperCamelCase_ ) if __name__ == "__main__": print("Number of terms ", solution(int(str(input()).strip())))
100
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor __magic_name__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ ( __a ): """simple docstring""" def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__): warnings.warn( """The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use YolosImageProcessor instead.""" , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__)
100
1
"""simple docstring""" import functools from typing import Any def a__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> bool: # Validation if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) or len(__SCREAMING_SNAKE_CASE ) == 0: raise ValueError("the string should be not empty string" ) if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) or not all( isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) and len(__SCREAMING_SNAKE_CASE ) > 0 for item in words ): raise ValueError("the words should be a list of non-empty strings" ) # Build trie __lowerCAmelCase: dict[str, Any] = {} __lowerCAmelCase: Dict = "WORD_KEEPER" for word in words: __lowerCAmelCase: Tuple = trie for c in word: if c not in trie_node: __lowerCAmelCase: str = {} __lowerCAmelCase: List[Any] = trie_node[c] __lowerCAmelCase: Dict = True __lowerCAmelCase: Any = len(__SCREAMING_SNAKE_CASE ) # Dynamic programming method @functools.cache def is_breakable(__SCREAMING_SNAKE_CASE ) -> bool: if index == len_string: return True __lowerCAmelCase: Optional[Any] = trie for i in range(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): __lowerCAmelCase: Dict = trie_node.get(string[i] , __SCREAMING_SNAKE_CASE ) if trie_node is None: return False if trie_node.get(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
108
"""simple docstring""" import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class snake_case ( __snake_case, unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Tuple = PriorTransformer SCREAMING_SNAKE_CASE_ : List[str] = """hidden_states""" @property def lowercase_ ( self : Dict)-> str: '''simple docstring''' __lowerCAmelCase: str = 4 __lowerCAmelCase: int = 8 __lowerCAmelCase: int = 7 __lowerCAmelCase: str = floats_tensor((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: Optional[Any] = floats_tensor((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: Any = floats_tensor((batch_size, num_embeddings, embedding_dim)).to(UpperCamelCase__) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowercase_ ( self : Optional[int] , UpperCamelCase__ : str=0)-> str: '''simple docstring''' torch.manual_seed(UpperCamelCase__) __lowerCAmelCase: List[Any] = 4 __lowerCAmelCase: Dict = 8 __lowerCAmelCase: int = 7 __lowerCAmelCase: List[str] = torch.randn((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: Tuple = torch.randn((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: List[Any] = torch.randn((batch_size, num_embeddings, embedding_dim)).to(UpperCamelCase__) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def lowercase_ ( self : Dict)-> List[Any]: '''simple docstring''' return (4, 8) @property def lowercase_ ( self : Optional[int])-> int: '''simple docstring''' return (4, 8) def lowercase_ ( self : Optional[int])-> Tuple: '''simple docstring''' __lowerCAmelCase: str = { "num_attention_heads": 2, "attention_head_dim": 4, "num_layers": 2, "embedding_dim": 8, "num_embeddings": 7, "additional_embeddings": 4, } __lowerCAmelCase: Any = self.dummy_input return init_dict, inputs_dict def lowercase_ ( self : List[Any])-> int: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase: Optional[int] = PriorTransformer.from_pretrained( "hf-internal-testing/prior-dummy" , output_loading_info=UpperCamelCase__) self.assertIsNotNone(UpperCamelCase__) self.assertEqual(len(loading_info["missing_keys"]) , 0) model.to(UpperCamelCase__) __lowerCAmelCase: Dict = model(**self.dummy_input)[0] assert hidden_states is not None, "Make sure output is not None" def lowercase_ ( self : List[str])-> Tuple: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase: Optional[Any] = self.prepare_init_args_and_inputs_for_common() __lowerCAmelCase: Tuple = self.model_class(**UpperCamelCase__) __lowerCAmelCase: List[str] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase: List[Any] = [*signature.parameters.keys()] __lowerCAmelCase: Any = ["hidden_states", "timestep"] self.assertListEqual(arg_names[:2] , UpperCamelCase__) def lowercase_ ( self : Optional[int])-> List[str]: '''simple docstring''' __lowerCAmelCase: int = PriorTransformer.from_pretrained("hf-internal-testing/prior-dummy") __lowerCAmelCase: Union[str, Any] = model.to(UpperCamelCase__) if hasattr(UpperCamelCase__ , "set_default_attn_processor"): model.set_default_attn_processor() __lowerCAmelCase: str = self.get_dummy_seed_input() with torch.no_grad(): __lowerCAmelCase: Dict = model(**UpperCamelCase__)[0] __lowerCAmelCase: Dict = output[0, :5].flatten().cpu() print(UpperCamelCase__) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. __lowerCAmelCase: List[str] = torch.tensor([-1.3436, -0.2870, 0.7538, 0.4368, -0.0239]) self.assertTrue(torch_all_close(UpperCamelCase__ , UpperCamelCase__ , rtol=1e-2)) @slow class snake_case ( unittest.TestCase ): def lowercase_ ( self : int , UpperCamelCase__ : Dict=1 , UpperCamelCase__ : str=7_6_8 , UpperCamelCase__ : int=7_7 , UpperCamelCase__ : Any=0)-> Union[str, Any]: '''simple docstring''' torch.manual_seed(UpperCamelCase__) __lowerCAmelCase: List[Any] = batch_size __lowerCAmelCase: Any = embedding_dim __lowerCAmelCase: Dict = num_embeddings __lowerCAmelCase: Dict = torch.randn((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: str = torch.randn((batch_size, embedding_dim)).to(UpperCamelCase__) __lowerCAmelCase: int = torch.randn((batch_size, num_embeddings, embedding_dim)).to(UpperCamelCase__) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowercase_ ( self : List[Any])-> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [1_3, [-0.5861, 0.1283, -0.0931, 0.0882, 0.4476, 0.1329, -0.0498, 0.0640]], [3_7, [-0.4913, 0.0110, -0.0483, 0.0541, 0.4954, -0.0170, 0.0354, 0.1651]], # fmt: on ]) def lowercase_ ( self : Optional[int] , UpperCamelCase__ : List[str] , UpperCamelCase__ : int)-> List[Any]: '''simple docstring''' __lowerCAmelCase: List[str] = PriorTransformer.from_pretrained("kandinsky-community/kandinsky-2-1-prior" , subfolder="prior") model.to(UpperCamelCase__) __lowerCAmelCase: Dict = self.get_dummy_seed_input(seed=UpperCamelCase__) with torch.no_grad(): __lowerCAmelCase: Optional[Any] = model(**UpperCamelCase__)[0] assert list(sample.shape) == [1, 7_6_8] __lowerCAmelCase: Dict = sample[0, :8].flatten().cpu() print(UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = torch.tensor(UpperCamelCase__) assert torch_all_close(UpperCamelCase__ , UpperCamelCase__ , atol=1e-3)
108
1
'''simple docstring''' def __lowerCamelCase ( lowerCAmelCase_ ) -> str: _a : Optional[Any] = '' for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def __lowerCamelCase ( lowerCAmelCase_ ) -> dict[str, str]: _a : List[str] = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key _a : int = remove_duplicates(key.upper() ) _a : str = len(lowerCAmelCase_ ) # First fill cipher with key characters _a : Dict = {alphabet[i]: char for i, char in enumerate(lowerCAmelCase_ )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(lowerCAmelCase_ ) , 26 ): _a : Optional[int] = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 _a : Tuple = alphabet[i - offset] _a : Optional[Any] = char return cipher_alphabet def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> str: return "".join(cipher_map.get(lowerCAmelCase_ , lowerCAmelCase_ ) for ch in message.upper() ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _a : Dict = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(lowerCAmelCase_ , lowerCAmelCase_ ) for ch in message.upper() ) def __lowerCamelCase ( ) -> None: _a : Optional[int] = input('Enter message to encode or decode: ' ).strip() _a : Dict = input('Enter keyword: ' ).strip() _a : Optional[int] = input('Encipher or decipher? E/D:' ).strip()[0].lower() try: _a : List[str] = {'e': encipher, 'd': decipher}[option] except KeyError: raise KeyError('invalid input option' ) _a : Dict = create_cipher_map(lowerCAmelCase_ ) print(func(lowerCAmelCase_ , lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
89
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __UpperCamelCase : Dict = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE( a_ ): _UpperCAmelCase = ["pixel_values"] def __init__( self: List[Any] , UpperCamelCase: bool = True , UpperCamelCase: Optional[Dict[str, int]] = None , UpperCamelCase: PILImageResampling = PILImageResampling.BILINEAR , UpperCamelCase: bool = True , UpperCamelCase: Dict[str, int] = None , UpperCamelCase: bool = True , UpperCamelCase: Union[int, float] = 1 / 2_55 , UpperCamelCase: bool = True , UpperCamelCase: Optional[Union[float, List[float]]] = None , UpperCamelCase: Optional[Union[float, List[float]]] = None , **UpperCamelCase: Optional[int] , ) -> None: super().__init__(**UpperCamelCase ) snake_case__ = size if size is not None else {'shortest_edge': 2_56} snake_case__ = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase ) snake_case__ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} snake_case__ = get_size_dict(UpperCamelCase ) snake_case__ = do_resize snake_case__ = size snake_case__ = resample snake_case__ = do_center_crop snake_case__ = crop_size snake_case__ = do_rescale snake_case__ = rescale_factor snake_case__ = do_normalize snake_case__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCAmelCase_ ( self: Tuple , UpperCamelCase: np.ndarray , UpperCamelCase: Dict[str, int] , UpperCamelCase: PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase: Dict , ) -> np.ndarray: snake_case__ = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) snake_case__ = get_resize_output_image_size(UpperCamelCase , size=size['shortest_edge'] , default_to_square=UpperCamelCase ) return resize(UpperCamelCase , size=UpperCamelCase , resample=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCAmelCase_ ( self: List[Any] , UpperCamelCase: np.ndarray , UpperCamelCase: Dict[str, int] , UpperCamelCase: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase: List[Any] , ) -> np.ndarray: snake_case__ = get_size_dict(UpperCamelCase ) return center_crop(UpperCamelCase , size=(size['height'], size['width']) , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCAmelCase_ ( self: Union[str, Any] , UpperCamelCase: np.ndarray , UpperCamelCase: float , UpperCamelCase: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase: Dict ) -> np.ndarray: return rescale(UpperCamelCase , scale=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCAmelCase_ ( self: Optional[Any] , UpperCamelCase: np.ndarray , UpperCamelCase: Union[float, List[float]] , UpperCamelCase: Union[float, List[float]] , UpperCamelCase: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase: Any , ) -> np.ndarray: return normalize(UpperCamelCase , mean=UpperCamelCase , std=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCAmelCase_ ( self: Any , UpperCamelCase: ImageInput , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Dict[str, int] = None , UpperCamelCase: PILImageResampling = None , UpperCamelCase: bool = None , UpperCamelCase: Dict[str, int] = None , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Optional[float] = None , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Optional[Union[float, List[float]]] = None , UpperCamelCase: Optional[Union[float, List[float]]] = None , UpperCamelCase: Optional[Union[str, TensorType]] = None , UpperCamelCase: Union[str, ChannelDimension] = ChannelDimension.FIRST , **UpperCamelCase: Any , ) -> Optional[Any]: snake_case__ = do_resize if do_resize is not None else self.do_resize snake_case__ = size if size is not None else self.size snake_case__ = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase ) snake_case__ = resample if resample is not None else self.resample snake_case__ = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case__ = crop_size if crop_size is not None else self.crop_size snake_case__ = get_size_dict(UpperCamelCase ) snake_case__ = do_rescale if do_rescale is not None else self.do_rescale snake_case__ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case__ = do_normalize if do_normalize is not None else self.do_normalize snake_case__ = image_mean if image_mean is not None else self.image_mean snake_case__ = image_std if image_std is not None else self.image_std snake_case__ = make_list_of_images(UpperCamelCase ) if not valid_images(UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. snake_case__ = [to_numpy_array(UpperCamelCase ) for image in images] if do_resize: snake_case__ = [self.resize(image=UpperCamelCase , size=UpperCamelCase , resample=UpperCamelCase ) for image in images] if do_center_crop: snake_case__ = [self.center_crop(image=UpperCamelCase , size=UpperCamelCase ) for image in images] if do_rescale: snake_case__ = [self.rescale(image=UpperCamelCase , scale=UpperCamelCase ) for image in images] if do_normalize: snake_case__ = [self.normalize(image=UpperCamelCase , mean=UpperCamelCase , std=UpperCamelCase ) for image in images] snake_case__ = [to_channel_dimension_format(UpperCamelCase , UpperCamelCase ) for image in images] snake_case__ = {'pixel_values': images} return BatchFeature(data=UpperCamelCase , tensor_type=UpperCamelCase )
307
0
import cva import numpy as np class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase , lowercase ) -> Optional[Any]: if k in (0.0_4, 0.0_6): lowerCamelCase_ = k lowerCamelCase_ = window_size else: raise ValueError("invalid k value" ) def __str__( self ) -> str: return str(self.k ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> tuple[cva.Mat, list[list[int]]]: lowerCamelCase_ = cva.imread(lowercase , 0 ) lowerCamelCase_ , lowerCamelCase_ = img.shape lowerCamelCase_ = [] lowerCamelCase_ = img.copy() lowerCamelCase_ = cva.cvtColor(lowercase , cva.COLOR_GRAY2RGB ) lowerCamelCase_ , lowerCamelCase_ = np.gradient(lowercase ) lowerCamelCase_ = dx**2 lowerCamelCase_ = dy**2 lowerCamelCase_ = dx * dy lowerCamelCase_ = 0.0_4 lowerCamelCase_ = self.window_size // 2 for y in range(lowercase , h - offset ): for x in range(lowercase , w - offset ): lowerCamelCase_ = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() lowerCamelCase_ = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() lowerCamelCase_ = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() lowerCamelCase_ = (wxx * wyy) - (wxy**2) lowerCamelCase_ = wxx + wyy lowerCamelCase_ = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": __A =HarrisCorner(0.04, 3) __A, __A =edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
47
import copy import re class _SCREAMING_SNAKE_CASE : lowerCAmelCase__ = 'hp' lowerCAmelCase__ = {} lowerCAmelCase__ = None @classmethod def SCREAMING_SNAKE_CASE_( cls , lowercase , lowercase ) -> Tuple: lowerCamelCase_ = prefix lowerCamelCase_ = defaults cls.build_naming_info() @staticmethod def SCREAMING_SNAKE_CASE_( lowercase , lowercase ) -> Optional[Any]: if len(lowercase ) == 0: return "" lowerCamelCase_ = None if any(char.isdigit() for char in word ): raise Exception(f'Parameters should not contain numbers: \'{word}\' contains a number' ) if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(lowercase ) + 1 ): lowerCamelCase_ = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: lowerCamelCase_ = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(lowercase ): lowerCamelCase_ = "" while integer != 0: lowerCamelCase_ = chr(ord("A" ) + integer % 10 ) + s integer //= 10 return s lowerCamelCase_ = 0 while True: lowerCamelCase_ = word + "#" + int_to_alphabetic(lowercase ) if sword in info["reverse_short_word"]: continue else: lowerCamelCase_ = sword break lowerCamelCase_ = short_word lowerCamelCase_ = word return short_word @staticmethod def SCREAMING_SNAKE_CASE_( lowercase , lowercase ) -> int: lowerCamelCase_ = param_name.split("_" ) lowerCamelCase_ = [TrialShortNamer.shortname_for_word(lowercase , lowercase ) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name lowerCamelCase_ = ["", "_"] for separator in separators: lowerCamelCase_ = separator.join(lowercase ) if shortname not in info["reverse_short_param"]: lowerCamelCase_ = shortname lowerCamelCase_ = param_name return shortname return param_name @staticmethod def SCREAMING_SNAKE_CASE_( lowercase , lowercase ) -> Optional[Any]: lowerCamelCase_ = TrialShortNamer.shortname_for_key(lowercase , lowercase ) lowerCamelCase_ = short_name lowerCamelCase_ = param_name @classmethod def SCREAMING_SNAKE_CASE_( cls ) -> Dict: if cls.NAMING_INFO is not None: return lowerCamelCase_ = { "short_word": {}, "reverse_short_word": {}, "short_param": {}, "reverse_short_param": {}, } lowerCamelCase_ = list(cls.DEFAULTS.keys() ) for k in field_keys: cls.add_new_param_name(lowercase , lowercase ) lowerCamelCase_ = info @classmethod def SCREAMING_SNAKE_CASE_( cls , lowercase ) -> Optional[int]: cls.build_naming_info() assert cls.PREFIX is not None lowerCamelCase_ = [copy.copy(cls.PREFIX )] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(f'You should provide a default value for the param name {k} with value {v}' ) if v == cls.DEFAULTS[k]: # The default value is not added to the name continue lowerCamelCase_ = cls.NAMING_INFO["short_param"][k] if isinstance(lowercase , lowercase ): lowerCamelCase_ = 1 if v else 0 lowerCamelCase_ = "" if isinstance(lowercase , (int, float) ) else "-" lowerCamelCase_ = f'{key}{sep}{v}' name.append(lowercase ) return "_".join(lowercase ) @classmethod def SCREAMING_SNAKE_CASE_( cls , lowercase ) -> List[Any]: lowerCamelCase_ = repr[len(cls.PREFIX ) + 1 :] if repr == "": lowerCamelCase_ = [] else: lowerCamelCase_ = repr.split("_" ) lowerCamelCase_ = {} for value in values: if "-" in value: lowerCamelCase_ , lowerCamelCase_ = value.split("-" ) else: lowerCamelCase_ = re.sub("[0-9.]" , "" , lowercase ) lowerCamelCase_ = float(re.sub("[^0-9.]" , "" , lowercase ) ) lowerCamelCase_ = cls.NAMING_INFO["reverse_short_param"][p_k] lowerCamelCase_ = p_v for k in cls.DEFAULTS: if k not in parameters: lowerCamelCase_ = cls.DEFAULTS[k] return parameters
47
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") UpperCAmelCase_ : Dict = logging.getLogger(__name__) @dataclass class lowerCAmelCase__ : '''simple docstring''' __UpperCamelCase = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) __UpperCamelCase = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) @dataclass class lowerCAmelCase__ : '''simple docstring''' __UpperCamelCase = field(default=UpperCAmelCase__ , metadata={"help": "The input training data file (a text file)."} ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "Overwrite the cached training and evaluation sets"} ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={"help": "The number of processes to use for the preprocessing."} , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={ "help": ( "The maximum total input sequence length after tokenization. If passed, sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={ "help": ( "Whether to pad all samples to the maximum sentence length. " "If False, will pad the samples dynamically when batching to the maximum length in the batch. More " "efficient on GPU but very bad for TPU." ) } , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) __UpperCamelCase = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' if self.train_file is not None: SCREAMING_SNAKE_CASE_ : List[Any] = self.train_file.split('''.''')[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: SCREAMING_SNAKE_CASE_ : List[str] = self.validation_file.split('''.''')[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowerCAmelCase__ : '''simple docstring''' __UpperCamelCase = 42 __UpperCamelCase = True __UpperCamelCase = None __UpperCamelCase = None def __call__( self : List[str] , lowercase_ : Any): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = '''label''' if '''label''' in features[0].keys() else '''labels''' SCREAMING_SNAKE_CASE_ : Union[str, Any] = [feature.pop(lowercase_) for feature in features] SCREAMING_SNAKE_CASE_ : Any = len(lowercase_) SCREAMING_SNAKE_CASE_ : Optional[Any] = len(features[0]['''input_ids''']) SCREAMING_SNAKE_CASE_ : Tuple = [ [{k: v[i] for k, v in feature.items()} for i in range(lowercase_)] for feature in features ] SCREAMING_SNAKE_CASE_ : Union[str, Any] = list(chain(*lowercase_)) SCREAMING_SNAKE_CASE_ : Dict = self.tokenizer.pad( lowercase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten SCREAMING_SNAKE_CASE_ : int = {k: v.view(lowercase_ , lowercase_ , -1) for k, v in batch.items()} # Add back labels SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.tensor(lowercase_ , dtype=torch.intaa) return batch def _A () -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[Any] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , __a , __a ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() SCREAMING_SNAKE_CASE_ : Optional[int] = training_args.get_process_log_level() logger.setLevel(__a ) datasets.utils.logging.set_verbosity(__a ) transformers.utils.logging.set_verbosity(__a ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. SCREAMING_SNAKE_CASE_ : Dict = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: SCREAMING_SNAKE_CASE_ : Tuple = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: SCREAMING_SNAKE_CASE_ : Tuple = {} if data_args.train_file is not None: SCREAMING_SNAKE_CASE_ : int = data_args.train_file if data_args.validation_file is not None: SCREAMING_SNAKE_CASE_ : Optional[Any] = data_args.validation_file SCREAMING_SNAKE_CASE_ : List[Any] = data_args.train_file.split('''.''' )[-1] SCREAMING_SNAKE_CASE_ : Optional[Any] = load_dataset( __a , data_files=__a , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. SCREAMING_SNAKE_CASE_ : List[str] = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. SCREAMING_SNAKE_CASE_ : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) SCREAMING_SNAKE_CASE_ : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) SCREAMING_SNAKE_CASE_ : str = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__a , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. SCREAMING_SNAKE_CASE_ : List[str] = [f'ending{i}' for i in range(4 )] SCREAMING_SNAKE_CASE_ : int = '''sent1''' SCREAMING_SNAKE_CASE_ : Optional[Any] = '''sent2''' if data_args.max_seq_length is None: SCREAMING_SNAKE_CASE_ : Any = tokenizer.model_max_length if max_seq_length > 10_24: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) SCREAMING_SNAKE_CASE_ : Tuple = 10_24 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) SCREAMING_SNAKE_CASE_ : int = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(__a ): SCREAMING_SNAKE_CASE_ : int = [[context] * 4 for context in examples[context_name]] SCREAMING_SNAKE_CASE_ : Any = examples[question_header_name] SCREAMING_SNAKE_CASE_ : Union[str, Any] = [ [f'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(__a ) ] # Flatten out SCREAMING_SNAKE_CASE_ : int = list(chain(*__a ) ) SCREAMING_SNAKE_CASE_ : List[Any] = list(chain(*__a ) ) # Tokenize SCREAMING_SNAKE_CASE_ : List[str] = tokenizer( __a , __a , truncation=__a , max_length=__a , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(__a ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) SCREAMING_SNAKE_CASE_ : int = raw_datasets['''train'''] if data_args.max_train_samples is not None: SCREAMING_SNAKE_CASE_ : Tuple = min(len(__a ) , data_args.max_train_samples ) SCREAMING_SNAKE_CASE_ : int = train_dataset.select(range(__a ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): SCREAMING_SNAKE_CASE_ : Optional[Any] = train_dataset.map( __a , batched=__a , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) SCREAMING_SNAKE_CASE_ : List[Any] = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: SCREAMING_SNAKE_CASE_ : Optional[int] = min(len(__a ) , data_args.max_eval_samples ) SCREAMING_SNAKE_CASE_ : List[Any] = eval_dataset.select(range(__a ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): SCREAMING_SNAKE_CASE_ : Optional[int] = eval_dataset.map( __a , batched=__a , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator SCREAMING_SNAKE_CASE_ : str = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=__a , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(__a ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[int] = eval_predictions SCREAMING_SNAKE_CASE_ : int = np.argmax(__a , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer SCREAMING_SNAKE_CASE_ : Dict = Trainer( model=__a , args=__a , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=__a , data_collator=__a , compute_metrics=__a , ) # Training if training_args.do_train: SCREAMING_SNAKE_CASE_ : int = None if training_args.resume_from_checkpoint is not None: SCREAMING_SNAKE_CASE_ : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: SCREAMING_SNAKE_CASE_ : List[str] = last_checkpoint SCREAMING_SNAKE_CASE_ : List[str] = trainer.train(resume_from_checkpoint=__a ) trainer.save_model() # Saves the tokenizer too for easy upload SCREAMING_SNAKE_CASE_ : Optional[Any] = train_result.metrics SCREAMING_SNAKE_CASE_ : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(__a ) ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = min(__a , len(__a ) ) trainer.log_metrics('''train''' , __a ) trainer.save_metrics('''train''' , __a ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) SCREAMING_SNAKE_CASE_ : Any = trainer.evaluate() SCREAMING_SNAKE_CASE_ : Any = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(__a ) SCREAMING_SNAKE_CASE_ : List[str] = min(__a , len(__a ) ) trainer.log_metrics('''eval''' , __a ) trainer.save_metrics('''eval''' , __a ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**__a ) else: trainer.create_model_card(**__a ) def _A (__a ) -> Any: """simple docstring""" main() if __name__ == "__main__": main()
91
"""simple docstring""" from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCAmelCase_ : Optional[int] = logging.get_logger(__name__) def _A (__a , __a ) -> Tuple: """simple docstring""" try: with open(__a , '''rb''' ) as flax_state_f: SCREAMING_SNAKE_CASE_ : Optional[int] = from_bytes(__a , flax_state_f.read() ) except UnpicklingError as e: try: with open(__a ) as f: if f.read().startswith('''version''' ): raise OSError( '''You seem to have cloned a repository without having git-lfs installed. Please''' ''' install git-lfs and run `git lfs install` followed by `git lfs pull` in the''' ''' folder you cloned.''' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f'Unable to convert {model_file} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(__a , __a ) def _A (__a , __a ) -> Tuple: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( '''Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights SCREAMING_SNAKE_CASE_ : Optional[int] = flatten_dict(jax.tree_util.tree_map(lambda __a : x.dtype == jnp.bfloataa , __a ) ).values() if any(__a ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = jax.tree_util.tree_map( lambda __a : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , __a ) SCREAMING_SNAKE_CASE_ : int = '''''' SCREAMING_SNAKE_CASE_ : str = flatten_dict(__a , sep='''.''' ) SCREAMING_SNAKE_CASE_ : List[Any] = pt_model.state_dict() # keep track of unexpected & missing keys SCREAMING_SNAKE_CASE_ : str = [] SCREAMING_SNAKE_CASE_ : Any = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): SCREAMING_SNAKE_CASE_ : Any = flax_key_tuple.split('''.''' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: SCREAMING_SNAKE_CASE_ : Any = flax_key_tuple_array[:-1] + ['''weight'''] SCREAMING_SNAKE_CASE_ : Optional[Any] = jnp.transpose(__a , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": SCREAMING_SNAKE_CASE_ : Tuple = flax_key_tuple_array[:-1] + ['''weight'''] SCREAMING_SNAKE_CASE_ : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": SCREAMING_SNAKE_CASE_ : Optional[int] = flax_key_tuple_array[:-1] + ['''weight'''] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(__a ): SCREAMING_SNAKE_CASE_ : List[str] = ( flax_key_tuple_string.replace('''_0''' , '''.0''' ) .replace('''_1''' , '''.1''' ) .replace('''_2''' , '''.2''' ) .replace('''_3''' , '''.3''' ) .replace('''_4''' , '''.4''' ) .replace('''_5''' , '''.5''' ) .replace('''_6''' , '''.6''' ) .replace('''_7''' , '''.7''' ) .replace('''_8''' , '''.8''' ) .replace('''_9''' , '''.9''' ) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = '''.'''.join(__a ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' f'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict SCREAMING_SNAKE_CASE_ : Optional[int] = np.asarray(__a ) if not isinstance(__a , np.ndarray ) else flax_tensor SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.from_numpy(__a ) # remove from missing keys missing_keys.remove(__a ) else: # weight is not expected by PyTorch model unexpected_keys.append(__a ) pt_model.load_state_dict(__a ) # re-transform missing_keys to list SCREAMING_SNAKE_CASE_ : int = list(__a ) if len(__a ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' f' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' f' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' f' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) if len(__a ) > 0: logger.warning( f'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' f' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ''' use it for predictions and inference.''' ) return pt_model
91
1
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __magic_name__ = TypeVar("KT") __magic_name__ = TypeVar("VT") class lowercase ( Generic[KT, VT] ): '''simple docstring''' def __init__( self , _snake_case = "root" , _snake_case = None ) -> List[Any]: """simple docstring""" UpperCAmelCase = key UpperCAmelCase = value UpperCAmelCase = [] def __repr__( self ) -> Optional[Any]: """simple docstring""" return f"""Node({self.key}: {self.value})""" @property def snake_case_ ( self ) -> Optional[int]: """simple docstring""" return len(self.forward ) class lowercase ( Generic[KT, VT] ): '''simple docstring''' def __init__( self , _snake_case = 0.5 , _snake_case = 16 ) -> Dict: """simple docstring""" UpperCAmelCase = Node[KT, VT]() UpperCAmelCase = 0 UpperCAmelCase = p UpperCAmelCase = max_level def __str__( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = list(self ) if len(__SCREAMING_SNAKE_CASE ) == 0: return f"""SkipList(level={self.level})""" UpperCAmelCase = max((len(str(__SCREAMING_SNAKE_CASE ) ) for item in items) , default=4 ) UpperCAmelCase = max(__SCREAMING_SNAKE_CASE , 4 ) + 4 UpperCAmelCase = self.head UpperCAmelCase = [] UpperCAmelCase = node.forward.copy() lines.append(f"""[{node.key}]""".ljust(__SCREAMING_SNAKE_CASE , '''-''' ) + '''* ''' * len(__SCREAMING_SNAKE_CASE ) ) lines.append(''' ''' * label_size + '''| ''' * len(__SCREAMING_SNAKE_CASE ) ) while len(node.forward ) != 0: UpperCAmelCase = node.forward[0] lines.append( f"""[{node.key}]""".ljust(__SCREAMING_SNAKE_CASE , '''-''' ) + ''' '''.join(str(n.key ) if n.key == node.key else '''|''' for n in forwards ) ) lines.append(''' ''' * label_size + '''| ''' * len(__SCREAMING_SNAKE_CASE ) ) UpperCAmelCase = node.forward lines.append('''None'''.ljust(__SCREAMING_SNAKE_CASE ) + '''* ''' * len(__SCREAMING_SNAKE_CASE ) ) return f"""SkipList(level={self.level})\n""" + "\n".join(__SCREAMING_SNAKE_CASE ) def __iter__( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCAmelCase = node.forward[0] def snake_case_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = 1 while random() < self.p and level < self.max_level: level += 1 return level def snake_case_ ( self , _snake_case ) -> Dict: """simple docstring""" UpperCAmelCase = [] UpperCAmelCase = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCAmelCase = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(__SCREAMING_SNAKE_CASE ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def snake_case_ ( self , _snake_case ) -> Any: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self._locate_node(__SCREAMING_SNAKE_CASE ) if node is not None: for i, update_node in enumerate(__SCREAMING_SNAKE_CASE ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCAmelCase = node.forward[i] else: UpperCAmelCase = update_node.forward[:i] def snake_case_ ( self , _snake_case , _snake_case ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self._locate_node(__SCREAMING_SNAKE_CASE ) if node is not None: UpperCAmelCase = value else: UpperCAmelCase = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , __SCREAMING_SNAKE_CASE ): update_vector.append(self.head ) UpperCAmelCase = level UpperCAmelCase = Node(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase = new_node def snake_case_ ( self , _snake_case ) -> Optional[int]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self._locate_node(__SCREAMING_SNAKE_CASE ) if node is not None: return node.value return None def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key1''' , 3 ) skip_list.insert('''Key2''' , 12 ) skip_list.insert('''Key3''' , 41 ) skip_list.insert('''Key4''' , -19 ) UpperCAmelCase = skip_list.head UpperCAmelCase = {} while node.level != 0: UpperCAmelCase = node.forward[0] UpperCAmelCase = node.value assert len(lowercase__ ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key1''' , 10 ) skip_list.insert('''Key1''' , 12 ) skip_list.insert('''Key5''' , 7 ) skip_list.insert('''Key7''' , 10 ) skip_list.insert('''Key10''' , 5 ) skip_list.insert('''Key7''' , 7 ) skip_list.insert('''Key5''' , 5 ) skip_list.insert('''Key10''' , 10 ) UpperCAmelCase = skip_list.head UpperCAmelCase = {} while node.level != 0: UpperCAmelCase = node.forward[0] UpperCAmelCase = node.value if len(lowercase__ ) != 4: print() assert len(lowercase__ ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() assert skip_list.find('''Some key''' ) is None def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key2''' , 20 ) assert skip_list.find('''Key2''' ) == 20 skip_list.insert('''Some Key''' , 10 ) skip_list.insert('''Key2''' , 8 ) skip_list.insert('''V''' , 13 ) assert skip_list.find('''Y''' ) is None assert skip_list.find('''Key2''' ) == 8 assert skip_list.find('''Some Key''' ) == 10 assert skip_list.find('''V''' ) == 13 def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.delete('''Some key''' ) assert len(skip_list.head.forward ) == 0 def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''Key2''' ) is None def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) == 14 assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''X''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key1''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) is None def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 142 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''X''' ) def traverse_keys(A__: Optional[int] ): yield node.key for forward_node in node.forward: yield from traverse_keys(lowercase__ ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def _lowerCAmelCase ( ): '''simple docstring''' def is_sorted(A__: Optional[int] ): return all(next_item >= item for item, next_item in zip(lowercase__ , lst[1:] ) ) UpperCAmelCase = SkipList() for i in range(10 ): skip_list.insert(lowercase__ , lowercase__ ) assert is_sorted(list(lowercase__ ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(lowercase__ ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(lowercase__ ) ) def _lowerCAmelCase ( ): '''simple docstring''' for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = SkipList() skip_list.insert(2 , '''2''' ) skip_list.insert(4 , '''4''' ) skip_list.insert(6 , '''4''' ) skip_list.insert(4 , '''5''' ) skip_list.insert(8 , '''4''' ) skip_list.insert(9 , '''4''' ) skip_list.delete(4 ) print(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
363
import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) __magic_name__ = logging.getLogger() def _lowerCAmelCase ( A__: Path , A__: list ): '''simple docstring''' UpperCAmelCase = '''\n'''.join(A__ ) Path(A__ ).open('''w''' ).writelines(A__ ) __magic_name__ = "patrickvonplaten/t5-tiny-random" __magic_name__ = "sshleifer/bart-tiny-random" __magic_name__ = "sshleifer/tiny-mbart" __magic_name__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowercase ( A__ ): '''simple docstring''' def snake_case_ ( self , _snake_case ) -> int: """simple docstring""" UpperCAmelCase = Path(self.get_auto_remove_tmp_dir() ) / '''utest_input.source''' UpperCAmelCase = input_file_name.parent / '''utest_output.txt''' assert not output_file_name.exists() UpperCAmelCase = [''' New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County.'''] _dump_articles(_snake_case , _snake_case ) UpperCAmelCase = str(Path(self.get_auto_remove_tmp_dir() ) / '''scores.json''' ) UpperCAmelCase = '''translation_en_to_de''' if model == T5_TINY else '''summarization''' UpperCAmelCase = f""" run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 """.split() with patch.object(_snake_case , '''argv''' , _snake_case ): run_generate() assert Path(_snake_case ).exists() # os.remove(Path(output_file_name)) def snake_case_ ( self ) -> Dict: """simple docstring""" self.run_eval_tester(_snake_case ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def snake_case_ ( self , _snake_case ) -> Any: """simple docstring""" self.run_eval_tester(_snake_case ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def snake_case_ ( self , _snake_case ) -> Optional[int]: """simple docstring""" UpperCAmelCase = Path(self.get_auto_remove_tmp_dir() ) / '''utest_input.source''' UpperCAmelCase = input_file_name.parent / '''utest_output.txt''' assert not output_file_name.exists() UpperCAmelCase = { '''en''': ['''Machine learning is great, isn\'t it?''', '''I like to eat bananas''', '''Tomorrow is another great day!'''], '''de''': [ '''Maschinelles Lernen ist großartig, oder?''', '''Ich esse gerne Bananen''', '''Morgen ist wieder ein toller Tag!''', ], } UpperCAmelCase = Path(self.get_auto_remove_tmp_dir() ) UpperCAmelCase = str(tmp_dir / '''scores.json''' ) UpperCAmelCase = str(tmp_dir / '''val.target''' ) _dump_articles(_snake_case , text['''en'''] ) _dump_articles(_snake_case , text['''de'''] ) UpperCAmelCase = '''translation_en_to_de''' if model == T5_TINY else '''summarization''' UpperCAmelCase = f""" run_eval_search.py {model} {str(_snake_case )} {str(_snake_case )} --score_path {score_path} --reference_path {reference_path} --task {task} """.split() testargs.extend(['''--search''', '''num_beams=1:2 length_penalty=0.9:1.0'''] ) with patch.object(_snake_case , '''argv''' , _snake_case ): with CaptureStdout() as cs: run_search() UpperCAmelCase = [''' num_beams | length_penalty''', model, '''Best score args'''] UpperCAmelCase = ['''Info'''] if "translation" in task: expected_strings.append('''bleu''' ) else: expected_strings.extend(_snake_case ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(_snake_case ).exists() os.remove(Path(_snake_case ) )
152
0
"""simple docstring""" import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _a = JukeboxTokenizer _a = { "artist": "Zac Brown Band", "genres": "Country", "lyrics": "I met a traveller from an antique land,\n Who said \"Two vast and trunkless legs of stone\n Stand in the desert. . . . Near them, on the sand,\n Half sunk a shattered visage lies, whose frown,\n And wrinkled lip, and sneer of cold command,\n Tell that its sculptor well those passions read\n Which yet survive, stamped on these lifeless things,\n The hand that mocked them, and the heart that fed;\n And on the pedestal, these words appear:\n My name is Ozymandias, King of Kings;\n Look on my Works, ye Mighty, and despair!\n Nothing beside remains. Round the decay\n Of that colossal Wreck, boundless and bare\n The lone and level sands stretch far away\n ", } @require_torch def __lowercase ( self : Union[str, Any] ): import torch lowerCAmelCase = JukeboxTokenizer.from_pretrained("""openai/jukebox-1b-lyrics""" ) lowerCAmelCase = tokenizer(**self.metas )["""input_ids"""] # fmt: off lowerCAmelCase = [ torch.tensor([[ 0, 0, 0, 7169, 507, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def __lowercase ( self : Optional[Any] ): import torch lowerCAmelCase = JukeboxTokenizer.from_pretrained("""openai/jukebox-5b-lyrics""" ) lowerCAmelCase = tokenizer(**self.metas )["""input_ids"""] # fmt: off lowerCAmelCase = [ torch.tensor([[ 0, 0, 0, 1069, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
155
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _lowercase : Tuple = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _lowercase : List[str] = 25_00_04 _lowercase : int = 25_00_20 @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( _lowerCAmelCase , unittest.TestCase ): a__ : Union[str, Any] = MBartaaTokenizer a__ : List[str] = MBartaaTokenizerFast a__ : Any = True a__ : List[str] = True def a ( self : str ): super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase = MBartaaTokenizer(_lowercase , src_lang='''en_XX''' , tgt_lang='''ro_RO''' , keep_accents=_lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def a ( self : Dict ): __UpperCAmelCase = '''<s>''' __UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowercase ) , _lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowercase ) , _lowercase ) def a ( self : Optional[Any] ): __UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(_lowercase ) , 10_54 ) def a ( self : Tuple ): self.assertEqual(self.get_tokenizer().vocab_size , 10_54 ) def a ( self : str ): __UpperCAmelCase = MBartaaTokenizer(_lowercase , src_lang='''en_XX''' , tgt_lang='''ro_RO''' , keep_accents=_lowercase ) __UpperCAmelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowercase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __UpperCAmelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowercase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.'''] , ) __UpperCAmelCase = tokenizer.convert_tokens_to_ids(_lowercase ) self.assertListEqual( _lowercase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(_lowercase ) self.assertListEqual( _lowercase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.'''] , ) @slow def a ( self : str ): # fmt: off __UpperCAmelCase = {'''input_ids''': [[25_00_04, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [25_00_04, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_00_04, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowercase , model_name='''facebook/mbart-large-50''' , revision='''d3913889c59cd5c9e456b269c376325eabad57e2''' , ) def a ( self : str ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart50''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __UpperCAmelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(_lowercase , _lowercase ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=True __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase , legacy_format=_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files self.assertSequenceEqual(_lowercase , _lowercase ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=False __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase , legacy_format=_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) shutil.rmtree(_lowercase ) @require_torch @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( unittest.TestCase ): a__ : str = "facebook/mbart-large-50-one-to-many-mmt" a__ : Union[str, Any] = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] a__ : Any = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] a__ : Any = [EN_CODE, 8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2] @classmethod def a ( cls : Tuple ): __UpperCAmelCase = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) __UpperCAmelCase = 1 return cls def a ( self : Union[str, Any] ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 25_00_01 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 25_00_04 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 25_00_20 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''mr_IN'''] , 25_00_38 ) def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _lowercase ) def a ( self : Optional[Any] ): self.assertIn(_lowercase , self.tokenizer.all_special_ids ) __UpperCAmelCase = [RO_CODE, 8_84, 90_19, 96, 9, 9_16, 8_67_92, 36, 1_87_43, 1_55_96, 5, 2] __UpperCAmelCase = self.tokenizer.decode(_lowercase , skip_special_tokens=_lowercase ) __UpperCAmelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertNotIn(self.tokenizer.eos_token , _lowercase ) def a ( self : Optional[Any] ): __UpperCAmelCase = ['''this is gunna be a long sentence ''' * 20] assert isinstance(src_text[0] , _lowercase ) __UpperCAmelCase = 10 __UpperCAmelCase = self.tokenizer(_lowercase , max_length=_lowercase , truncation=_lowercase ).input_ids[0] self.assertEqual(ids[0] , _lowercase ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(_lowercase ) , _lowercase ) def a ( self : Optional[int] ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [25_00_53, 25_00_01] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowercase ) __UpperCAmelCase = MBartaaTokenizer.from_pretrained(_lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _lowercase ) @require_torch def a ( self : Dict ): __UpperCAmelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_lowercase , return_tensors='''pt''' ) __UpperCAmelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_lowercase , truncation=_lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) __UpperCAmelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __UpperCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer(self.src_text , padding=_lowercase , truncation=_lowercase , max_length=3 , return_tensors='''pt''' ) __UpperCAmelCase = self.tokenizer( text_target=self.tgt_text , padding=_lowercase , truncation=_lowercase , max_length=10 , return_tensors='''pt''' ) __UpperCAmelCase = targets['''input_ids'''] __UpperCAmelCase = shift_tokens_right(_lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def a ( self : Dict ): __UpperCAmelCase = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(_lowercase ) , { # en_XX, A, test, EOS '''input_ids''': [[25_00_04, 62, 30_34, 2]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 25_00_01, } , )
332
0
"""simple docstring""" def lowerCamelCase ( _UpperCamelCase : str ) -> str: '''simple docstring''' return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
366
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowerCamelCase__ ( metaclass=A ): """simple docstring""" __a = ["""keras_nlp"""] def __init__( self : str , *UpperCamelCase : List[Any] , **UpperCamelCase : Dict ): '''simple docstring''' requires_backends(self , ["""keras_nlp"""] )
320
0
"""simple docstring""" import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin SCREAMING_SNAKE_CASE_ = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class lowerCAmelCase_ : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=16 , snake_case_=13 , snake_case_=7 , snake_case_=14 , snake_case_=10 , snake_case_=19 , snake_case_=5 , snake_case_=4 , snake_case_=True , snake_case_=16 , snake_case_=2 , snake_case_=4 , snake_case_=4 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=[1, 2, 3, 4, 5] , snake_case_=25 , snake_case_=5 , ) -> Tuple: __lowerCAmelCase = d_model __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = prediction_length __lowerCAmelCase = context_length __lowerCAmelCase = cardinality __lowerCAmelCase = num_time_features __lowerCAmelCase = lags_sequence __lowerCAmelCase = embedding_dimension __lowerCAmelCase = is_training __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = context_length __lowerCAmelCase = prediction_length + label_length __lowerCAmelCase = label_length __lowerCAmelCase = moving_average __lowerCAmelCase = autocorrelation_factor def A__ ( self ) -> List[Any]: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def A__ ( self , snake_case_ ) -> Any: __lowerCAmelCase = config.context_length + max(config.lags_sequence ) __lowerCAmelCase = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length] ) __lowerCAmelCase = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.get_config() __lowerCAmelCase = self.prepare_autoformer_inputs_dict(snake_case_ ) return config, inputs_dict def A__ ( self ) -> int: __lowerCAmelCase , __lowerCAmelCase = self.prepare_config_and_inputs() return config, inputs_dict def A__ ( self , snake_case_ , snake_case_ ) -> int: __lowerCAmelCase = AutoformerModel(config=snake_case_ ).to(snake_case_ ).eval() __lowerCAmelCase = model(**snake_case_ ) __lowerCAmelCase = outputs.encoder_last_hidden_state __lowerCAmelCase = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_encoder() encoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerEncoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = model.create_network_inputs(**snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) __lowerCAmelCase = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) __lowerCAmelCase = encoder(inputs_embeds=snake_case_ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) __lowerCAmelCase = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) __lowerCAmelCase = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) __lowerCAmelCase = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) __lowerCAmelCase = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_decoder() decoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerDecoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase = decoder( trend=snake_case_ , inputs_embeds=snake_case_ , encoder_hidden_states=snake_case_ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class lowerCAmelCase_ ( A__ , A__ , unittest.TestCase ): '''simple docstring''' _snake_case = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () _snake_case = (AutoformerForPrediction,) if is_torch_available() else () _snake_case = {'''feature-extraction''': AutoformerModel} if is_torch_available() else {} _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False def A__ ( self ) -> Optional[int]: __lowerCAmelCase = AutoformerModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ ) def A__ ( self ) -> Optional[int]: self.config_tester.run_common_tests() def A__ ( self ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model_class.from_pretrained(snake_case_ , output_loading_info=snake_case_ ) self.assertEqual(info["""missing_keys"""] , [] ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*snake_case_ ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def A__ ( self ) -> Any: pass def A__ ( self ) -> str: __lowerCAmelCase = inspect.signature(getattr(snake_case_ , """forward""" ) ) # The main input is the name of the argument after `self` __lowerCAmelCase = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , snake_case_ ) def A__ ( self ) -> Any: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = True __lowerCAmelCase = getattr(self.model_tester , """seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """decoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """d_model""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """num_attention_heads""" , snake_case_ ) __lowerCAmelCase = d_model // num_attention_heads for model_class in self.all_model_classes: __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) __lowerCAmelCase = len(snake_case_ ) __lowerCAmelCase = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(snake_case_ , snake_case_ ) # decoder attentions __lowerCAmelCase = outputs.decoder_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions __lowerCAmelCase = outputs.cross_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine __lowerCAmelCase = True __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + 2 , len(snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def A__ ( self ) -> int: super().test_retain_grad_hidden_states_attentions() def lowercase (_lowerCAmelCase="train-batch.pt" ): __lowerCAmelCase = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=_lowerCAmelCase , repo_type="""dataset""" ) __lowerCAmelCase = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase ) return batch @require_torch @slow class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def A__ ( self ) -> int: __lowerCAmelCase = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch() with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , future_values=batch["""future_values"""] , future_time_features=batch["""future_time_features"""] , )[0] __lowerCAmelCase = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[0.3_593, -1.3_398, 0.6_330], [0.2_279, 1.5_396, -0.1_792], [0.0_450, 1.3_225, -0.2_335]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> List[str]: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , ).encoder_last_hidden_state __lowerCAmelCase = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[-0.0_734, -0.9_036, 0.8_358], [4.7_186, 2.4_113, 1.9_581], [1.7_953, 2.3_558, 1.2_970]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> Any: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model.generate( static_categorical_features=batch["""static_categorical_features"""] , past_time_features=batch["""past_time_features"""] , past_values=batch["""past_values"""] , future_time_features=batch["""future_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , ) __lowerCAmelCase = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , snake_case_ ) __lowerCAmelCase = torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=snake_case_ ) __lowerCAmelCase = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , snake_case_ , rtol=1e-1 ) )
301
"""simple docstring""" import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def lowercase (_lowerCAmelCase , _lowerCAmelCase="shi-labs/oneformer_demo" ): with open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) as f: __lowerCAmelCase = json.load(_lowerCAmelCase ) __lowerCAmelCase = {} __lowerCAmelCase = [] __lowerCAmelCase = [] for key, info in class_info.items(): __lowerCAmelCase = info["""name"""] class_names.append(info["""name"""] ) if info["isthing"]: thing_ids.append(int(_lowerCAmelCase ) ) __lowerCAmelCase = thing_ids __lowerCAmelCase = class_names return metadata class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=30 , snake_case_=400 , snake_case_=None , snake_case_=True , snake_case_=True , snake_case_=[0.5, 0.5, 0.5] , snake_case_=[0.5, 0.5, 0.5] , snake_case_=10 , snake_case_=False , snake_case_=255 , snake_case_="shi-labs/oneformer_demo" , snake_case_="ade20k_panoptic.json" , snake_case_=10 , ) -> Union[str, Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = {"""shortest_edge""": 32, """longest_edge""": 1_333} if size is None else size __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean __lowerCAmelCase = image_std __lowerCAmelCase = class_info_file __lowerCAmelCase = prepare_metadata(snake_case_ , snake_case_ ) __lowerCAmelCase = num_text __lowerCAmelCase = repo_path # for the post_process_functions __lowerCAmelCase = 2 __lowerCAmelCase = 10 __lowerCAmelCase = 10 __lowerCAmelCase = 3 __lowerCAmelCase = 4 __lowerCAmelCase = num_labels __lowerCAmelCase = do_reduce_labels __lowerCAmelCase = ignore_index def A__ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def A__ ( self , snake_case_ , snake_case_=False ) -> Dict: if not batched: __lowerCAmelCase = image_inputs[0] if isinstance(snake_case_ , Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image.size else: __lowerCAmelCase , __lowerCAmelCase = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase = int(self.size["""shortest_edge"""] * h / w ) __lowerCAmelCase = self.size["""shortest_edge"""] elif w > h: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = int(self.size["""shortest_edge"""] * w / h ) else: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = self.size["""shortest_edge"""] else: __lowerCAmelCase = [] for image in image_inputs: __lowerCAmelCase , __lowerCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[0] )[0] __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[1] )[1] return expected_height, expected_width def A__ ( self ) -> Tuple: return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class lowerCAmelCase_ ( A__ , unittest.TestCase ): '''simple docstring''' _snake_case = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string _snake_case = image_processing_class def A__ ( self ) -> str: __lowerCAmelCase = OneFormerImageProcessorTester(self ) @property def A__ ( self ) -> Dict: return self.image_processing_tester.prepare_image_processor_dict() def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case_ , """image_std""" ) ) self.assertTrue(hasattr(snake_case_ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case_ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case_ , """size""" ) ) self.assertTrue(hasattr(snake_case_ , """ignore_index""" ) ) self.assertTrue(hasattr(snake_case_ , """class_info_file""" ) ) self.assertTrue(hasattr(snake_case_ , """num_text""" ) ) self.assertTrue(hasattr(snake_case_ , """repo_path""" ) ) self.assertTrue(hasattr(snake_case_ , """metadata""" ) ) self.assertTrue(hasattr(snake_case_ , """do_reduce_labels""" ) ) def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Union[str, Any]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> List[str]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> Tuple: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self , snake_case_=False , snake_case_=False , snake_case_="np" ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # prepare image and target __lowerCAmelCase = self.image_processing_tester.num_labels __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) if with_segmentation_maps: __lowerCAmelCase = num_labels if is_instance_map: __lowerCAmelCase = list(range(snake_case_ ) ) * 2 __lowerCAmelCase = dict(enumerate(snake_case_ ) ) __lowerCAmelCase = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": __lowerCAmelCase = [Image.fromarray(snake_case_ ) for annotation in annotations] __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , snake_case_ , return_tensors="""pt""" , instance_id_to_semantic_id=snake_case_ , pad_and_return_pixel_mask=snake_case_ , ) return inputs def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Optional[Any]: def common(snake_case_=False , snake_case_=None ): __lowerCAmelCase = self.comm_get_image_processor_inputs( with_segmentation_maps=snake_case_ , is_instance_map=snake_case_ , segmentation_type=snake_case_ ) __lowerCAmelCase = inputs["""mask_labels"""] __lowerCAmelCase = inputs["""class_labels"""] __lowerCAmelCase = inputs["""pixel_values"""] __lowerCAmelCase = inputs["""text_inputs"""] # check the batch_size for mask_label, class_label, text_input in zip(snake_case_ , snake_case_ , snake_case_ ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.num_text ) common() common(is_instance_map=snake_case_ ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase = np.zeros((20, 50) ) __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = binary_mask_to_rle(snake_case_ ) self.assertEqual(len(snake_case_ ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) __lowerCAmelCase = [(1, 4) for i in range(self.image_processing_tester.batch_size )] __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ , target_sizes=snake_case_ ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_instance_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_panoptic_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
301
1
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[int]: lowerCAmelCase__ : Optional[int] = [True] * limit lowerCAmelCase__ : Optional[Any] = False lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : Dict = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): lowerCAmelCase__ : int = i * 2 while index < limit: lowerCAmelCase__ : List[str] = False lowerCAmelCase__ : List[Any] = index + i lowerCAmelCase__ : List[Any] = [2] for i in range(3 , SCREAMING_SNAKE_CASE_ , 2 ): if is_prime[i]: primes.append(SCREAMING_SNAKE_CASE_ ) return primes def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 1_000_000 ) -> int: lowerCAmelCase__ : List[Any] = prime_sieve(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : Optional[int] = 0 for i in range(len(SCREAMING_SNAKE_CASE_ ) ): for j in range(i + length , len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Union[str, Any] = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: lowerCAmelCase__ : int = j - i lowerCAmelCase__ : Optional[Any] = sol return largest if __name__ == "__main__": print(F"""{solution() = }""")
307
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: stooge(SCREAMING_SNAKE_CASE_ , 0 , len(SCREAMING_SNAKE_CASE_ ) - 1 ) return arr def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Dict: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: lowerCAmelCase__ , lowerCAmelCase__ : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: lowerCAmelCase__ : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , (h - t) ) # Recursively sort last 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , i + t , (SCREAMING_SNAKE_CASE_) ) # Recursively sort first 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , (h - t) ) if __name__ == "__main__": lowerCamelCase__ = input("""Enter numbers separated by a comma:\n""").strip() lowerCamelCase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
307
1
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case_ (A__ ): def __init__( self :int ,__snake_case :List[Any] ,__snake_case :List[Any] ) -> Any: super().__init__() # make sure scheduler can always be converted to DDIM a__ = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=__lowercase ,scheduler=__lowercase ) @torch.no_grad() def __call__( self :List[str] ,__snake_case :Optional[int] = 1 ,__snake_case :List[str] = None ,__snake_case :Any = 0.0 ,__snake_case :Tuple = 50 ,__snake_case :Dict = None ,__snake_case :Dict = "pil" ,__snake_case :Dict = True ,) -> Union[ImagePipelineOutput, Tuple]: if isinstance(self.unet.config.sample_size ,__lowercase ): a__ = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: a__ = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(__lowercase ,__lowercase ) and len(__lowercase ) != batch_size: raise ValueError( F'You have passed a list of generators of length {len(__lowercase )}, but requested an effective batch' F' size of {batch_size}. Make sure the batch size matches the length of the generators.' ) a__ = randn_tensor(__lowercase ,generator=__lowercase ,device=self.device ,dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output a__ = self.unet(__lowercase ,__lowercase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__ = self.scheduler.step( __lowercase ,__lowercase ,__lowercase ,eta=__lowercase ,use_clipped_model_output=__lowercase ,generator=__lowercase ).prev_sample a__ = (image / 2 + 0.5).clamp(0 ,1 ) a__ = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": a__ = self.numpy_to_pil(__lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__lowercase )
240
import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def lowerCAmelCase_ ( _lowercase : Dict , _lowercase : str , _lowercase : str , _lowercase : Optional[Any]=1024) -> List[Any]: """simple docstring""" a__ , a__ : Optional[int] = [], [] a__ : Union[str, Any] = list(zip(_lowercase , _lowercase)) a__ , a__ : List[Any] = sorted_examples[0] def is_too_big(_lowercase : Tuple): return tok(_lowercase , return_tensors="""pt""").input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:]): a__ : Tuple = new_src + """ """ + src a__ : Any = new_tgt + """ """ + tgt if is_too_big(_lowercase) or is_too_big(_lowercase): # cant fit, finalize example finished_src.append(_lowercase) finished_tgt.append(_lowercase) a__ , a__ : List[Any] = src, tgt else: # can fit, keep adding a__ , a__ : Tuple = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(_lowercase) finished_tgt.append(_lowercase) return finished_src, finished_tgt def lowerCAmelCase_ ( _lowercase : str , _lowercase : Path , _lowercase : Any , _lowercase : str) -> Tuple: """simple docstring""" a__ : Any = Path(_lowercase) save_path.mkdir(exist_ok=_lowercase) for split in ["train"]: a__ , a__ : List[Any] = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' a__ : Dict = [x.rstrip() for x in Path(_lowercase).open().readlines()] a__ : Optional[Any] = [x.rstrip() for x in Path(_lowercase).open().readlines()] a__ , a__ : List[Any] = pack_examples(_lowercase , _lowercase , _lowercase , _lowercase) print(F'''packed {split} split from {len(_lowercase)} examples -> {len(_lowercase)}.''') Path(save_path / F'''{split}.source''').open("""w""").write("""\n""".join(_lowercase)) Path(save_path / F'''{split}.target''').open("""w""").write("""\n""".join(_lowercase)) for split in ["val", "test"]: a__ , a__ : Any = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' shutil.copyfile(_lowercase , save_path / F'''{split}.source''') shutil.copyfile(_lowercase , save_path / F'''{split}.target''') def lowerCAmelCase_ ( ) -> Optional[int]: """simple docstring""" a__ : Tuple = argparse.ArgumentParser() parser.add_argument("""--tok_name""" , type=_lowercase , help="""like facebook/bart-large-cnn,t5-base, etc.""") parser.add_argument("""--max_seq_len""" , type=_lowercase , default=128) parser.add_argument("""--data_dir""" , type=_lowercase) parser.add_argument("""--save_path""" , type=_lowercase) a__ : List[Any] = parser.parse_args() a__ : List[Any] = AutoTokenizer.from_pretrained(args.tok_name) return pack_data_dir(_lowercase , Path(args.data_dir) , args.max_seq_len , args.save_path) if __name__ == "__main__": packer_cli()
170
0
from math import pow def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count __a = int(pow(_UpperCAmelCase , _UpperCAmelCase ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n __a = backtrack( _UpperCAmelCase , _UpperCAmelCase , current_number + 1 , _UpperCAmelCase , _UpperCAmelCase ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. __a = backtrack( _UpperCAmelCase , _UpperCAmelCase , current_number + 1 , _UpperCAmelCase , _UpperCAmelCase ) return current_sum, solutions_count def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): if not (1 <= needed_sum <= 1000 and 2 <= power <= 10): raise ValueError( '''Invalid input\n''' '''needed_sum must be between 1 and 1000, power between 2 and 10.''' ) return backtrack(_UpperCAmelCase , _UpperCAmelCase , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
370
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __snake_case :Optional[Any] = logging.get_logger(__name__) __snake_case :List[Any] = '''▁''' __snake_case :List[Any] = {'''vocab_file''': '''spiece.model'''} __snake_case :Tuple = { '''vocab_file''': { '''google/reformer-crime-and-punishment''': ( '''https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model''' ) } } __snake_case :List[Any] = { '''google/reformer-crime-and-punishment''': 52_4288, } class _A ( __UpperCAmelCase ): UpperCamelCase__ : int = VOCAB_FILES_NAMES UpperCamelCase__ : str = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Optional[int] = ['''input_ids''', '''attention_mask'''] def __init__( self : Any , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[Any]="</s>" , __SCREAMING_SNAKE_CASE : List[Any]="<unk>" , __SCREAMING_SNAKE_CASE : Any=[] , __SCREAMING_SNAKE_CASE : Optional[Dict[str, Any]] = None , **__SCREAMING_SNAKE_CASE : List[Any] , ): '''simple docstring''' __a = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , additional_special_tokens=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) __a = vocab_file __a = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(__SCREAMING_SNAKE_CASE) @property def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' return self.sp_model.get_piece_size() def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = {self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : Dict): '''simple docstring''' __a = self.__dict__.copy() __a = None return state def __setstate__( self : Dict , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' __a = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): __a = {} __a = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def _lowerCamelCase ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : str): '''simple docstring''' return self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' return self.sp_model.piece_to_id(__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[str] , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' if index < self.sp_model.get_piece_size(): __a = self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE) return token def _lowerCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : Any): '''simple docstring''' __a = [] __a = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE) + token __a = [] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE) out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE) return out_string.strip() def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None): '''simple docstring''' if not os.path.isdir(__SCREAMING_SNAKE_CASE): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return __a = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(__SCREAMING_SNAKE_CASE) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE) elif not os.path.isfile(self.vocab_file): with open(__SCREAMING_SNAKE_CASE , '''wb''') as fi: __a = self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE) return (out_vocab_file,)
131
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue_model_parallelism.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, ] ) class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Dict: if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="""utf-8""" , check=__A , ) assert hasattr(self , """env""" ) def __lowerCAmelCase ( self , __A ) -> Any: # configuration for running training on smdistributed Model Parallel lowerCAmelCase_ :Union[str, Any] = { """enabled""": True, """processes_per_host""": 8, } lowerCAmelCase_ :Tuple = { """enabled""": True, """parameters""": { """microbatches""": 4, """placement_strategy""": """spread""", """pipeline""": """interleaved""", """optimize""": """speed""", """partitions""": 4, """ddp""": True, }, } lowerCAmelCase_ :Any = {"""smdistributed""": {"""modelparallel""": smp_options}, """mpi""": mpi_options} lowerCAmelCase_ :Any = """trainer""" if self.script == """run_glue.py""" else """smtrainer""" # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-{instance_count}-smp-{name_extension}""" , instance_count=__A , instance_type=self.instance_type , debugger_hook_config=__A , hyperparameters={ **self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path, """max_steps""": 500, } , metric_definitions=self.env.metric_definitions , distribution=__A , py_version="""py36""" , ) def __lowerCAmelCase ( self , __A ) -> List[Any]: TrainingJobAnalytics(__A ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(1,)] ) def __lowerCAmelCase ( self , __A ) -> List[str]: # create estimator lowerCAmelCase_ :Any = self.create_estimator(__A ) # run training estimator.fit() # result dataframe lowerCAmelCase_ :Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis lowerCAmelCase_ :List[str] = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) lowerCAmelCase_ :Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping lowerCAmelCase_ :Optional[int] = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , __A )
84
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A__ : List[Any] = logging.get_logger(__name__) A__ : List[Any] = { 'SCUT-DLVCLab/lilt-roberta-en-base': ( 'https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json' ), } class lowercase__ ( snake_case__ ): _UpperCAmelCase :int = "lilt" def __init__( self : Tuple , snake_case__ : List[str]=3_0522 , snake_case__ : Tuple=768 , snake_case__ : Optional[int]=12 , snake_case__ : str=12 , snake_case__ : List[str]=3072 , snake_case__ : Optional[Any]="gelu" , snake_case__ : Tuple=0.1 , snake_case__ : List[str]=0.1 , snake_case__ : str=512 , snake_case__ : int=2 , snake_case__ : List[Any]=0.02 , snake_case__ : Optional[Any]=1E-12 , snake_case__ : Tuple=0 , snake_case__ : List[Any]="absolute" , snake_case__ : Tuple=None , snake_case__ : str=4 , snake_case__ : str=1024 , **snake_case__ : Dict , ): super().__init__(pad_token_id=snake_case__ , **snake_case__ ) lowerCamelCase_ : Tuple =vocab_size lowerCamelCase_ : Tuple =hidden_size lowerCamelCase_ : List[Any] =num_hidden_layers lowerCamelCase_ : Dict =num_attention_heads lowerCamelCase_ : List[str] =hidden_act lowerCamelCase_ : str =intermediate_size lowerCamelCase_ : Tuple =hidden_dropout_prob lowerCamelCase_ : str =attention_probs_dropout_prob lowerCamelCase_ : Optional[Any] =max_position_embeddings lowerCamelCase_ : List[str] =type_vocab_size lowerCamelCase_ : str =initializer_range lowerCamelCase_ : Optional[Any] =layer_norm_eps lowerCamelCase_ : Optional[int] =position_embedding_type lowerCamelCase_ : Optional[int] =classifier_dropout lowerCamelCase_ : Optional[Any] =channel_shrink_ratio lowerCamelCase_ : Dict =max_ad_position_embeddings
364
"""simple docstring""" import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase__ ( snake_case__ ): def __init__( self : Tuple , snake_case__ : Optional[int] , snake_case__ : int=None , snake_case__ : Union[str, Any]=True , snake_case__ : Optional[int]=None , **snake_case__ : Optional[int] ): lowerCamelCase_ : Dict =parent lowerCamelCase_ : List[str] =config_class lowerCamelCase_ : Union[str, Any] =has_text_modality lowerCamelCase_ : Optional[int] =kwargs lowerCamelCase_ : List[str] =common_properties def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : List[str] =self.config_class(**self.inputs_dict ) lowerCamelCase_ : Any =( ["hidden_size", "num_attention_heads", "num_hidden_layers"] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["vocab_size"] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(snake_case__ , snake_case__ ) , msg=F"""`{prop}` does not exist""" ) # Test that config has the common properties as setter for idx, name in enumerate(snake_case__ ): try: setattr(snake_case__ , snake_case__ , snake_case__ ) self.parent.assertEqual( getattr(snake_case__ , snake_case__ ) , snake_case__ , msg=F"""`{name} value {idx} expected, but was {getattr(snake_case__ , snake_case__ )}""" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(snake_case__ ): try: lowerCamelCase_ : Dict =self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(snake_case__ , snake_case__ ) , snake_case__ , msg=F"""`{name} value {idx} expected, but was {getattr(snake_case__ , snake_case__ )}""" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : Tuple =self.config_class(**self.inputs_dict ) lowerCamelCase_ : Any =json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , snake_case__ ) def UpperCAmelCase__ ( self : int ): lowerCamelCase_ : Tuple =self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ : List[Any] =os.path.join(snake_case__ , "config.json" ) config_first.to_json_file(snake_case__ ) lowerCamelCase_ : Optional[int] =self.config_class.from_json_file(snake_case__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : Dict =self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(snake_case__ ) lowerCamelCase_ : Optional[int] =self.config_class.from_pretrained(snake_case__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase__ ( self : str ): lowerCamelCase_ : Dict =self.config_class(**self.inputs_dict ) lowerCamelCase_ : Dict ="test" with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ : str =os.path.join(snake_case__ , snake_case__ ) config_first.save_pretrained(snake_case__ ) lowerCamelCase_ : Optional[Any] =self.config_class.from_pretrained(snake_case__ , subfolder=snake_case__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : Optional[Any] =self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) lowerCamelCase_ : List[Any] =3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def UpperCAmelCase__ ( self : List[Any] ): if self.config_class.is_composition: return lowerCamelCase_ : Tuple =self.config_class() self.parent.assertIsNotNone(snake_case__ ) def UpperCAmelCase__ ( self : List[Any] ): lowerCamelCase_ : List[str] =copy.deepcopy(snake_case__ ) lowerCamelCase_ : Optional[int] =self.config_class(**snake_case__ ) lowerCamelCase_ : Union[str, Any] =[] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("torch_dtype", config.torch_dtype, torch.floataa) ) elif getattr(snake_case__ , snake_case__ ) != value: wrong_values.append((key, getattr(snake_case__ , snake_case__ ), value) ) if len(snake_case__ ) > 0: lowerCamelCase_ : Any ="\n".join([F"""- {v[0]}: got {v[1]} instead of {v[2]}""" for v in wrong_values] ) raise ValueError(F"""The following keys were not properly set in the config:\n{errors}""" ) def UpperCAmelCase__ ( self : int ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
209
0